mirror of
https://github.com/ARM-software/devlib.git
synced 2025-04-04 17:00:03 +01:00
Default modules are a recurrent source of errors as they fail to initialize (cgroups particularly) on any recent target. This leads to error in basically any workload-automation setup on Android 12 and above targets. Since modules can now be lazily loaded upon Target attribute access, there is no reason to preload those anymore.
3189 lines
118 KiB
Python
3189 lines
118 KiB
Python
# Copyright 2024 ARM Limited
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
import asyncio
|
|
from contextlib import contextmanager
|
|
import io
|
|
import base64
|
|
import functools
|
|
import gzip
|
|
import glob
|
|
import os
|
|
from operator import itemgetter
|
|
import re
|
|
import time
|
|
import logging
|
|
import posixpath
|
|
import subprocess
|
|
import tarfile
|
|
import tempfile
|
|
import threading
|
|
import uuid
|
|
import xml.dom.minidom
|
|
import copy
|
|
import inspect
|
|
import itertools
|
|
from collections import namedtuple, defaultdict
|
|
from past.builtins import long
|
|
from past.types import basestring
|
|
from numbers import Number
|
|
from shlex import quote
|
|
try:
|
|
from collections.abc import Mapping
|
|
except ImportError:
|
|
from collections import Mapping
|
|
|
|
from enum import Enum
|
|
from concurrent.futures import ThreadPoolExecutor
|
|
|
|
from devlib.host import LocalConnection, PACKAGE_BIN_DIRECTORY
|
|
from devlib.module import get_module, Module
|
|
from devlib.platform import Platform
|
|
from devlib.exception import (DevlibTransientError, TargetStableError,
|
|
TargetNotRespondingError, TimeoutError,
|
|
TargetTransientError, KernelConfigKeyError,
|
|
TargetError, HostError, TargetCalledProcessError)
|
|
from devlib.utils.ssh import SshConnection
|
|
from devlib.utils.android import AdbConnection, AndroidProperties, LogcatMonitor, adb_command, INTENT_FLAGS
|
|
from devlib.utils.misc import memoized, isiterable, convert_new_lines, groupby_value
|
|
from devlib.utils.misc import commonprefix, merge_lists
|
|
from devlib.utils.misc import ABI_MAP, get_cpu_name, ranges_to_list
|
|
from devlib.utils.misc import batch_contextmanager, tls_property, _BoundTLSProperty, nullcontext
|
|
from devlib.utils.misc import safe_extract
|
|
from devlib.utils.types import integer, boolean, bitmask, identifier, caseless_string, bytes_regex
|
|
import devlib.utils.asyn as asyn
|
|
|
|
|
|
FSTAB_ENTRY_REGEX = re.compile(r'(\S+) on (.+) type (\S+) \((\S+)\)')
|
|
ANDROID_SCREEN_STATE_REGEX = re.compile('(?:mPowerState|mScreenOn|mWakefulness|Display Power: state)=([0-9]+|true|false|ON|OFF|DOZE|Dozing|Asleep|Awake)',
|
|
re.IGNORECASE)
|
|
ANDROID_SCREEN_RESOLUTION_REGEX = re.compile(r'cur=(?P<width>\d+)x(?P<height>\d+)')
|
|
ANDROID_SCREEN_ROTATION_REGEX = re.compile(r'orientation=(?P<rotation>[0-3])')
|
|
DEFAULT_SHELL_PROMPT = re.compile(r'^.*(shell|root|juno)@?.*:[/~]\S* *[#$] ',
|
|
re.MULTILINE)
|
|
KVERSION_REGEX = re.compile(
|
|
r'(?P<version>\d+)(\.(?P<major>\d+)(\.(?P<minor>\d+))?(-rc(?P<rc>\d+))?)?(-android(?P<android_version>[0-9]+))?(-(?P<commits>\d+)-g(?P<sha1>[0-9a-fA-F]{7,}))?(-ab(?P<gki_abi>[0-9]+))?'
|
|
)
|
|
|
|
GOOGLE_DNS_SERVER_ADDRESS = '8.8.8.8'
|
|
|
|
|
|
installed_package_info = namedtuple('installed_package_info', 'apk_path package')
|
|
|
|
|
|
def call_conn(f):
|
|
"""
|
|
Decorator to be used on all :class:`devlib.target.Target` methods that
|
|
directly use a method of ``self.conn``.
|
|
|
|
This ensures that if a call to any of the decorated method occurs while
|
|
executing, a new connection will be created in order to avoid possible
|
|
deadlocks. This can happen if e.g. a target's method is called from
|
|
``__del__``, which could be executed by the garbage collector, interrupting
|
|
another call to a method of the connection instance.
|
|
|
|
.. note:: This decorator could be applied directly to all methods with a
|
|
metaclass or ``__init_subclass__`` but it could create issues when
|
|
passing target methods as callbacks to connections' methods.
|
|
"""
|
|
|
|
@functools.wraps(f)
|
|
def wrapper(self, *args, **kwargs):
|
|
reentered = self.conn.is_in_use
|
|
disconnect = False
|
|
try:
|
|
# If the connection was already in use we need to use a different
|
|
# instance to avoid reentrancy deadlocks. This can happen even in
|
|
# single threaded code via __del__ implementations that can be
|
|
# called at any point.
|
|
if reentered:
|
|
# Shallow copy so we can use another connection instance
|
|
_self = copy.copy(self)
|
|
_self.conn = _self.get_connection()
|
|
assert self.conn is not _self.conn
|
|
disconnect = True
|
|
else:
|
|
_self = self
|
|
return f(_self, *args, **kwargs)
|
|
finally:
|
|
if disconnect:
|
|
_self.disconnect()
|
|
|
|
return wrapper
|
|
|
|
|
|
class Target(object):
|
|
|
|
path = None
|
|
os = None
|
|
system_id = None
|
|
|
|
default_modules = []
|
|
|
|
@property
|
|
def core_names(self):
|
|
return self.platform.core_names
|
|
|
|
@property
|
|
def core_clusters(self):
|
|
return self.platform.core_clusters
|
|
|
|
@property
|
|
def big_core(self):
|
|
return self.platform.big_core
|
|
|
|
@property
|
|
def little_core(self):
|
|
return self.platform.little_core
|
|
|
|
@property
|
|
def is_connected(self):
|
|
return self.conn is not None
|
|
|
|
@property
|
|
def connected_as_root(self):
|
|
return self.conn and self.conn.connected_as_root
|
|
|
|
@property
|
|
def is_rooted(self):
|
|
if self._is_rooted is None:
|
|
try:
|
|
self.execute('ls /', timeout=5, as_root=True)
|
|
self._is_rooted = True
|
|
except(TargetError, TimeoutError):
|
|
self._is_rooted = False
|
|
|
|
return self._is_rooted or self.connected_as_root
|
|
|
|
@property
|
|
@memoized
|
|
def needs_su(self):
|
|
return not self.connected_as_root and self.is_rooted
|
|
|
|
@property
|
|
@memoized
|
|
def kernel_version(self):
|
|
return KernelVersion(self.execute('{} uname -r -v'.format(quote(self.busybox))).strip())
|
|
|
|
@property
|
|
def hostid(self):
|
|
return int(self.execute('{} hostid'.format(self.busybox)).strip(), 16)
|
|
|
|
@property
|
|
def hostname(self):
|
|
return self.execute('{} hostname'.format(self.busybox)).strip()
|
|
|
|
@property
|
|
def os_version(self): # pylint: disable=no-self-use
|
|
return {}
|
|
|
|
@property
|
|
def model(self):
|
|
return self.platform.model
|
|
|
|
@property
|
|
def abi(self): # pylint: disable=no-self-use
|
|
return None
|
|
|
|
@property
|
|
def supported_abi(self):
|
|
return [self.abi]
|
|
|
|
@property
|
|
@memoized
|
|
def cpuinfo(self):
|
|
return Cpuinfo(self.execute('cat /proc/cpuinfo'))
|
|
|
|
@property
|
|
@memoized
|
|
def number_of_cpus(self):
|
|
num_cpus = 0
|
|
corere = re.compile(r'^\s*cpu\d+\s*$')
|
|
output = self.execute('ls /sys/devices/system/cpu', as_root=self.is_rooted)
|
|
for entry in output.split():
|
|
if corere.match(entry):
|
|
num_cpus += 1
|
|
return num_cpus
|
|
|
|
@property
|
|
@memoized
|
|
def number_of_nodes(self):
|
|
cmd = 'cd /sys/devices/system/node && {busybox} find . -maxdepth 1'.format(busybox=quote(self.busybox))
|
|
try:
|
|
output = self.execute(cmd, as_root=self.is_rooted)
|
|
except TargetStableError:
|
|
return 1
|
|
else:
|
|
nodere = re.compile(r'^\./node\d+\s*$')
|
|
num_nodes = 0
|
|
for entry in output.splitlines():
|
|
if nodere.match(entry):
|
|
num_nodes += 1
|
|
return num_nodes
|
|
|
|
@property
|
|
@memoized
|
|
def list_nodes_cpus(self):
|
|
nodes_cpus = []
|
|
for node in range(self.number_of_nodes):
|
|
path = self.path.join('/sys/devices/system/node/node{}/cpulist'.format(node))
|
|
output = self.read_value(path)
|
|
nodes_cpus.append(ranges_to_list(output))
|
|
return nodes_cpus
|
|
|
|
@property
|
|
@memoized
|
|
def config(self):
|
|
try:
|
|
return KernelConfig(self.execute('zcat /proc/config.gz'))
|
|
except TargetStableError:
|
|
for path in ['/boot/config-$({} uname -r)'.format(self.busybox), '/boot/config']:
|
|
try:
|
|
return KernelConfig(self.execute('cat {}'.format(path)))
|
|
except TargetStableError:
|
|
pass
|
|
return KernelConfig('')
|
|
|
|
@property
|
|
@memoized
|
|
def user(self):
|
|
return self.getenv('USER')
|
|
|
|
@property
|
|
@memoized
|
|
def page_size_kb(self):
|
|
cmd = "cat /proc/self/smaps | {0} grep KernelPageSize | {0} head -n 1 | {0} awk '{{ print $2 }}'"
|
|
return int(self.execute(cmd.format(self.busybox)) or 0)
|
|
|
|
@property
|
|
def shutils(self):
|
|
if self._shutils is None:
|
|
self._setup_shutils()
|
|
return self._shutils
|
|
|
|
def is_running(self, comm):
|
|
cmd_ps = f'''{self.busybox} ps -A -T -o stat,comm'''
|
|
cmd_awk = f'''{self.busybox} awk 'BEGIN{{found=0}} {{state=$1; $1=""; if ($state != "Z" && $0 == " {comm}") {{found=1}}}} END {{print found}}' '''
|
|
result = self.execute(f"{cmd_ps} | {cmd_awk}")
|
|
return bool(int(result))
|
|
|
|
@tls_property
|
|
def _conn(self):
|
|
try:
|
|
return self._unused_conns.pop()
|
|
except KeyError:
|
|
return self.get_connection()
|
|
|
|
# Add a basic property that does not require calling to get the value
|
|
conn = _conn.basic_property
|
|
|
|
@tls_property
|
|
def _async_manager(self):
|
|
return asyn.AsyncManager()
|
|
|
|
# Add a basic property that does not require calling to get the value
|
|
async_manager = _async_manager.basic_property
|
|
|
|
def __init__(self,
|
|
connection_settings=None,
|
|
platform=None,
|
|
working_directory=None,
|
|
executables_directory=None,
|
|
connect=True,
|
|
modules=None,
|
|
load_default_modules=True,
|
|
shell_prompt=DEFAULT_SHELL_PROMPT,
|
|
conn_cls=None,
|
|
is_container=False,
|
|
max_async=50,
|
|
):
|
|
self._async_pool = None
|
|
self._async_pool_size = None
|
|
self._unused_conns = set()
|
|
|
|
self._is_rooted = None
|
|
self.connection_settings = connection_settings or {}
|
|
# Set self.platform: either it's given directly (by platform argument)
|
|
# or it's given in the connection_settings argument
|
|
# If neither, create default Platform()
|
|
if platform is None:
|
|
self.platform = self.connection_settings.get('platform', Platform())
|
|
else:
|
|
self.platform = platform
|
|
# Check if the user hasn't given two different platforms
|
|
if 'platform' in self.connection_settings:
|
|
if connection_settings['platform'] is not platform:
|
|
raise TargetStableError('Platform specified in connection_settings '
|
|
'({}) differs from that directly passed '
|
|
'({})!)'
|
|
.format(connection_settings['platform'],
|
|
self.platform))
|
|
self.connection_settings['platform'] = self.platform
|
|
self.working_directory = working_directory
|
|
self.executables_directory = executables_directory
|
|
self.load_default_modules = load_default_modules
|
|
self.shell_prompt = bytes_regex(shell_prompt)
|
|
self.conn_cls = conn_cls
|
|
self.is_container = is_container
|
|
self.logger = logging.getLogger(self.__class__.__name__)
|
|
self._installed_binaries = {}
|
|
self._installed_modules = {}
|
|
self._cache = {}
|
|
self._shutils = None
|
|
self._file_transfer_cache = None
|
|
self._max_async = max_async
|
|
self.busybox = None
|
|
|
|
def normalize_mod_spec(spec):
|
|
if isinstance(spec, str):
|
|
return (spec, {})
|
|
else:
|
|
[(name, params)] = spec.items()
|
|
return (name, params)
|
|
|
|
modules = sorted(
|
|
map(
|
|
normalize_mod_spec,
|
|
itertools.chain(
|
|
self.default_modules if load_default_modules else [],
|
|
modules or [],
|
|
self.platform.modules or [],
|
|
)
|
|
),
|
|
key=itemgetter(0),
|
|
)
|
|
|
|
# Ensure that we did not ask for the same module but different
|
|
# configurations. Empty configurations are ignored, so any
|
|
# user-provided conf will win against an empty conf.
|
|
def elect(name, specs):
|
|
specs = list(specs)
|
|
|
|
confs = set(
|
|
tuple(sorted(params.items()))
|
|
for _, params in specs
|
|
if params
|
|
)
|
|
if len(confs) > 1:
|
|
raise ValueError(f'Attempted to load the module "{name}" with multiple different configuration')
|
|
else:
|
|
if any(
|
|
params is None
|
|
for _, params in specs
|
|
):
|
|
params = None
|
|
else:
|
|
params = dict(confs.pop()) if confs else {}
|
|
|
|
return (name, params)
|
|
|
|
modules = dict(itertools.starmap(
|
|
elect,
|
|
itertools.groupby(modules, key=itemgetter(0))
|
|
))
|
|
|
|
def get_kind(name):
|
|
return get_module(name).kind or ''
|
|
|
|
def kind_conflict(kind, names):
|
|
if kind:
|
|
raise ValueError(f'Cannot enable multiple modules sharing the same kind "{kind}": {sorted(names)}')
|
|
|
|
list(itertools.starmap(
|
|
kind_conflict,
|
|
itertools.groupby(
|
|
sorted(
|
|
modules.keys(),
|
|
key=get_kind
|
|
),
|
|
key=get_kind
|
|
)
|
|
))
|
|
self._modules = modules
|
|
|
|
self._update_modules('early')
|
|
if connect:
|
|
self.connect(max_async=max_async)
|
|
|
|
def __getstate__(self):
|
|
# tls_property will recreate the underlying value automatically upon
|
|
# access and is typically used for dynamic content that cannot be
|
|
# pickled or should not transmitted to another thread.
|
|
ignored = {
|
|
k
|
|
for k, v in inspect.getmembers(self.__class__)
|
|
if isinstance(v, _BoundTLSProperty)
|
|
}
|
|
ignored.update((
|
|
'_async_pool',
|
|
'_unused_conns',
|
|
))
|
|
return {
|
|
k: v
|
|
for k, v in self.__dict__.items()
|
|
if k not in ignored
|
|
}
|
|
|
|
def __setstate__(self, dct):
|
|
self.__dict__ = dct
|
|
pool_size = self._async_pool_size
|
|
if pool_size is None:
|
|
self._async_pool = None
|
|
else:
|
|
self._async_pool = ThreadPoolExecutor(pool_size)
|
|
self._unused_conns = set()
|
|
|
|
# connection and initialization
|
|
|
|
@asyn.asyncf
|
|
async def connect(self, timeout=None, check_boot_completed=True, max_async=None):
|
|
self.platform.init_target_connection(self)
|
|
# Forcefully set the thread-local value for the connection, with the
|
|
# timeout we want
|
|
self.conn = self.get_connection(timeout=timeout)
|
|
if check_boot_completed:
|
|
self.wait_boot_complete(timeout)
|
|
self.check_connection()
|
|
self._resolve_paths()
|
|
self.execute('mkdir -p {}'.format(quote(self.working_directory)))
|
|
self.execute('mkdir -p {}'.format(quote(self.executables_directory)))
|
|
self.busybox = self.install(os.path.join(PACKAGE_BIN_DIRECTORY, self.abi, 'busybox'), timeout=30)
|
|
self.conn.busybox = self.busybox
|
|
self._detect_max_async(max_async or self._max_async)
|
|
self.platform.update_from_target(self)
|
|
self._update_modules('connected')
|
|
|
|
def _detect_max_async(self, max_async):
|
|
self.logger.debug('Detecting max number of async commands ...')
|
|
|
|
def make_conn(_):
|
|
try:
|
|
conn = self.get_connection()
|
|
except Exception:
|
|
return None
|
|
else:
|
|
payload = 'hello'
|
|
# Sanity check the connection, in case we managed to connect
|
|
# but it's actually unusable.
|
|
try:
|
|
res = conn.execute(f'echo {quote(payload)}')
|
|
except Exception:
|
|
return None
|
|
else:
|
|
if res.strip() == payload:
|
|
return conn
|
|
else:
|
|
return None
|
|
|
|
# Logging needs to be disabled before the thread pool is created,
|
|
# otherwise the logging config will not be taken into account
|
|
logging.disable()
|
|
try:
|
|
# Aggressively attempt to create all the connections in parallel,
|
|
# so that this setup step does not take too much time.
|
|
with ThreadPoolExecutor(max_async) as pool:
|
|
conns = pool.map(make_conn, range(max_async))
|
|
# Avoid polluting the log with errors coming from broken
|
|
# connections.
|
|
finally:
|
|
logging.disable(logging.NOTSET)
|
|
|
|
conns = {conn for conn in conns if conn is not None}
|
|
|
|
# Keep the connection so it can be reused by future threads
|
|
self._unused_conns.update(conns)
|
|
max_conns = len(conns)
|
|
|
|
self.logger.debug(f'Detected max number of async commands: {max_conns}')
|
|
self._async_pool_size = max_conns
|
|
self._async_pool = ThreadPoolExecutor(max_conns)
|
|
|
|
@asyn.asyncf
|
|
async def check_connection(self):
|
|
"""
|
|
Check that the connection works without obvious issues.
|
|
"""
|
|
out = await self.execute.asyn('true', as_root=False)
|
|
if out.strip():
|
|
raise TargetStableError('The shell seems to not be functional and adds content to stderr: {}'.format(out))
|
|
|
|
def disconnect(self):
|
|
connections = self._conn.get_all_values()
|
|
for conn in itertools.chain(connections, self._unused_conns):
|
|
conn.close()
|
|
if self._async_pool is not None:
|
|
self._async_pool.__exit__(None, None, None)
|
|
|
|
def get_connection(self, timeout=None):
|
|
if self.conn_cls is None:
|
|
raise ValueError('Connection class not specified on Target creation.')
|
|
conn = self.conn_cls(timeout=timeout, **self.connection_settings) # pylint: disable=not-callable
|
|
# This allows forwarding the detected busybox for connections created in new threads.
|
|
conn.busybox = self.busybox
|
|
return conn
|
|
|
|
def wait_boot_complete(self, timeout=10):
|
|
raise NotImplementedError()
|
|
|
|
@asyn.asyncf
|
|
async def setup(self, executables=None):
|
|
await self._setup_shutils.asyn()
|
|
|
|
for host_exe in (executables or []): # pylint: disable=superfluous-parens
|
|
await self.install.asyn(host_exe)
|
|
|
|
# Check for platform dependent setup procedures
|
|
self.platform.setup(self)
|
|
|
|
# Initialize modules which requires Busybox (e.g. shutil dependent tasks)
|
|
self._update_modules('setup')
|
|
|
|
await self.execute.asyn('mkdir -p {}'.format(quote(self._file_transfer_cache)))
|
|
|
|
def reboot(self, hard=False, connect=True, timeout=180):
|
|
if hard:
|
|
if not self.has('hard_reset'):
|
|
raise TargetStableError('Hard reset not supported for this target.')
|
|
self.hard_reset() # pylint: disable=no-member
|
|
else:
|
|
if not self.is_connected:
|
|
message = 'Cannot reboot target because it is disconnected. ' +\
|
|
'Either connect() first, or specify hard=True ' +\
|
|
'(in which case, a hard_reset module must be installed)'
|
|
raise TargetTransientError(message)
|
|
self.reset()
|
|
# Wait a fixed delay before starting polling to give the target time to
|
|
# shut down, otherwise, might create the connection while it's still shutting
|
|
# down resulting in subsequent connection failing.
|
|
self.logger.debug('Waiting for target to power down...')
|
|
reset_delay = 20
|
|
time.sleep(reset_delay)
|
|
timeout = max(timeout - reset_delay, 10)
|
|
if self.has('boot'):
|
|
self.boot() # pylint: disable=no-member
|
|
self.conn.connected_as_root = None
|
|
if connect:
|
|
self.connect(timeout=timeout)
|
|
|
|
# file transfer
|
|
|
|
@asyn.asynccontextmanager
|
|
async def _xfer_cache_path(self, name):
|
|
"""
|
|
Context manager to provide a unique path in the transfer cache with the
|
|
basename of the given name.
|
|
"""
|
|
# Use a UUID to avoid race conditions on the target side
|
|
xfer_uuid = uuid.uuid4().hex
|
|
folder = self.path.join(self._file_transfer_cache, xfer_uuid)
|
|
# Make sure basename will work on folders too
|
|
name = os.path.normpath(name)
|
|
# Ensure the name is relative so that os.path.join() will actually
|
|
# join the paths rather than ignoring the first one.
|
|
name = './{}'.format(os.path.basename(name))
|
|
|
|
check_rm = False
|
|
try:
|
|
await self.makedirs.asyn(folder)
|
|
# Don't check the exit code as the folder might not even exist
|
|
# before this point, if creating it failed
|
|
check_rm = True
|
|
yield self.path.join(folder, name)
|
|
finally:
|
|
await self.execute.asyn('rm -rf -- {}'.format(quote(folder)), check_exit_code=check_rm)
|
|
|
|
@asyn.asyncf
|
|
async def _prepare_xfer(self, action, sources, dest, pattern=None, as_root=False):
|
|
"""
|
|
Check the sanity of sources and destination and prepare the ground for
|
|
transfering multiple sources.
|
|
"""
|
|
|
|
once = functools.lru_cache(maxsize=None)
|
|
|
|
_target_cache = {}
|
|
def target_paths_kind(paths, as_root=False):
|
|
def process(x):
|
|
x = x.strip()
|
|
if x == 'notexist':
|
|
return None
|
|
else:
|
|
return x
|
|
|
|
_paths = [
|
|
path
|
|
for path in paths
|
|
if path not in _target_cache
|
|
]
|
|
if _paths:
|
|
cmd = '; '.join(
|
|
'if [ -d {path} ]; then echo dir; elif [ -e {path} ]; then echo file; else echo notexist; fi'.format(
|
|
path=quote(path)
|
|
)
|
|
for path in _paths
|
|
)
|
|
res = self.execute(cmd, as_root=as_root)
|
|
_target_cache.update(zip(_paths, map(process, res.split())))
|
|
|
|
return [
|
|
_target_cache[path]
|
|
for path in paths
|
|
]
|
|
|
|
_host_cache = {}
|
|
def host_paths_kind(paths, as_root=False):
|
|
def path_kind(path):
|
|
if os.path.isdir(path):
|
|
return 'dir'
|
|
elif os.path.exists(path):
|
|
return 'file'
|
|
else:
|
|
return None
|
|
|
|
for path in paths:
|
|
if path not in _host_cache:
|
|
_host_cache[path] = path_kind(path)
|
|
|
|
return [
|
|
_host_cache[path]
|
|
for path in paths
|
|
]
|
|
|
|
# TODO: Target.remove() and Target.makedirs() would probably benefit
|
|
# from being implemented by connections, with the current
|
|
# implementation in ConnectionBase. This would allow SshConnection to
|
|
# use SFTP for these operations, which should be cheaper than
|
|
# Target.execute()
|
|
if action == 'push':
|
|
src_excep = HostError
|
|
src_path_kind = host_paths_kind
|
|
|
|
_dst_mkdir = once(self.makedirs)
|
|
dst_path_join = self.path.join
|
|
dst_paths_kind = target_paths_kind
|
|
dst_remove_file = once(functools.partial(self.remove, as_root=as_root))
|
|
elif action == 'pull':
|
|
src_excep = TargetStableError
|
|
src_path_kind = target_paths_kind
|
|
|
|
_dst_mkdir = once(functools.partial(os.makedirs, exist_ok=True))
|
|
dst_path_join = os.path.join
|
|
dst_paths_kind = host_paths_kind
|
|
dst_remove_file = once(os.remove)
|
|
else:
|
|
raise ValueError('Unknown action "{}"'.format(action))
|
|
|
|
# Handle the case where path is None
|
|
def dst_mkdir(path):
|
|
if path:
|
|
_dst_mkdir(path)
|
|
|
|
def rewrite_dst(src, dst):
|
|
new_dst = dst_path_join(dst, os.path.basename(src))
|
|
|
|
src_kind, = src_path_kind([src], as_root)
|
|
# Batch both checks to avoid a costly extra execute()
|
|
dst_kind, new_dst_kind = dst_paths_kind([dst, new_dst], as_root)
|
|
|
|
if src_kind == 'file':
|
|
if dst_kind == 'dir':
|
|
if new_dst_kind == 'dir':
|
|
raise IsADirectoryError(new_dst)
|
|
if new_dst_kind == 'file':
|
|
dst_remove_file(new_dst)
|
|
return new_dst
|
|
else:
|
|
return new_dst
|
|
elif dst_kind == 'file':
|
|
dst_remove_file(dst)
|
|
return dst
|
|
else:
|
|
dst_mkdir(os.path.dirname(dst))
|
|
return dst
|
|
elif src_kind == 'dir':
|
|
if dst_kind == 'dir':
|
|
# Do not allow writing over an existing folder
|
|
if new_dst_kind == 'dir':
|
|
raise FileExistsError(new_dst)
|
|
if new_dst_kind == 'file':
|
|
raise FileExistsError(new_dst)
|
|
else:
|
|
return new_dst
|
|
elif dst_kind == 'file':
|
|
raise FileExistsError(dst_kind)
|
|
else:
|
|
dst_mkdir(os.path.dirname(dst))
|
|
return dst
|
|
else:
|
|
raise FileNotFoundError(src)
|
|
|
|
if pattern:
|
|
if not sources:
|
|
raise src_excep('No file matching source pattern: {}'.format(pattern))
|
|
|
|
if dst_paths_kind([dest]) != ['dir']:
|
|
raise NotADirectoryError('A folder dest is required for multiple matches but destination is a file: {}'.format(dest))
|
|
|
|
# TODO: since rewrite_dst() will currently return a different path for
|
|
# each source, it will not bring anything. In order to be useful,
|
|
# connections need to be able to understand that if the destination is
|
|
# an empty folder, the source is supposed to be transfered into it with
|
|
# the same basename.
|
|
return groupby_value({
|
|
src: rewrite_dst(src, dest)
|
|
for src in sources
|
|
})
|
|
|
|
@asyn.asyncf
|
|
@call_conn
|
|
async def push(self, source, dest, as_root=False, timeout=None, globbing=False): # pylint: disable=arguments-differ
|
|
source = str(source)
|
|
dest = str(dest)
|
|
|
|
sources = glob.glob(source) if globbing else [source]
|
|
mapping = await self._prepare_xfer.asyn('push', sources, dest, pattern=source if globbing else None, as_root=as_root)
|
|
|
|
def do_push(sources, dest):
|
|
for src in sources:
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='host', path=src, mode='r')
|
|
)
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=dest, mode='w')
|
|
)
|
|
return self.conn.push(sources, dest, timeout=timeout)
|
|
|
|
if as_root:
|
|
for sources, dest in mapping.items():
|
|
for source in sources:
|
|
async with self._xfer_cache_path(source) as device_tempfile:
|
|
do_push([source], device_tempfile)
|
|
await self.execute.asyn("mv -f -- {} {}".format(quote(device_tempfile), quote(dest)), as_root=True)
|
|
else:
|
|
for sources, dest in mapping.items():
|
|
do_push(sources, dest)
|
|
|
|
@asyn.asyncf
|
|
async def _expand_glob(self, pattern, **kwargs):
|
|
"""
|
|
Expand the given path globbing pattern on the target using the shell
|
|
globbing.
|
|
"""
|
|
# Since we split the results based on new lines, forbid them in the
|
|
# pattern
|
|
if '\n' in pattern:
|
|
raise ValueError(r'Newline character \n are not allowed in globbing patterns')
|
|
|
|
# If the pattern is in fact a plain filename, skip the expansion on the
|
|
# target to avoid an unncessary command execution.
|
|
#
|
|
# fnmatch char list from: https://docs.python.org/3/library/fnmatch.html
|
|
special_chars = ['*', '?', '[', ']']
|
|
if not any(char in pattern for char in special_chars):
|
|
return [pattern]
|
|
|
|
# Characters to escape that are impacting parameter splitting, since we
|
|
# want the pattern to be given in one piece. Unfortunately, there is no
|
|
# fool-proof way of doing that without also escaping globbing special
|
|
# characters such as wildcard which would defeat the entire purpose of
|
|
# that function.
|
|
for c in [' ', "'", '"']:
|
|
pattern = pattern.replace(c, '\\' + c)
|
|
|
|
cmd = "exec printf '%s\n' {}".format(pattern)
|
|
# Make sure to use the same shell everywhere for the path globbing,
|
|
# ensuring consistent results no matter what is the default platform
|
|
# shell
|
|
cmd = '{} sh -c {} 2>/dev/null'.format(quote(self.busybox), quote(cmd))
|
|
# On some shells, match failure will make the command "return" a
|
|
# non-zero code, even though the command was not actually called
|
|
result = await self.execute.asyn(cmd, strip_colors=False, check_exit_code=False, **kwargs)
|
|
paths = result.splitlines()
|
|
if not paths:
|
|
raise TargetStableError('No file matching: {}'.format(pattern))
|
|
|
|
return paths
|
|
|
|
@asyn.asyncf
|
|
@call_conn
|
|
async def pull(self, source, dest, as_root=False, timeout=None, globbing=False, via_temp=False): # pylint: disable=arguments-differ
|
|
source = str(source)
|
|
dest = str(dest)
|
|
|
|
if globbing:
|
|
sources = await self._expand_glob.asyn(source, as_root=as_root)
|
|
else:
|
|
sources = [source]
|
|
|
|
# The SSH server might not have the right permissions to read the file,
|
|
# so use a temporary copy instead.
|
|
via_temp |= as_root
|
|
|
|
mapping = await self._prepare_xfer.asyn('pull', sources, dest, pattern=source if globbing else None, as_root=as_root)
|
|
|
|
def do_pull(sources, dest):
|
|
for src in sources:
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=src, mode='r')
|
|
)
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='host', path=dest, mode='w')
|
|
)
|
|
self.conn.pull(sources, dest, timeout=timeout)
|
|
|
|
if via_temp:
|
|
for sources, dest in mapping.items():
|
|
for source in sources:
|
|
async with self._xfer_cache_path(source) as device_tempfile:
|
|
await self.execute.asyn("cp -r -- {} {}".format(quote(source), quote(device_tempfile)), as_root=as_root)
|
|
await self.execute.asyn("{} chmod 0644 -- {}".format(self.busybox, quote(device_tempfile)), as_root=as_root)
|
|
do_pull([device_tempfile], dest)
|
|
else:
|
|
for sources, dest in mapping.items():
|
|
do_pull(sources, dest)
|
|
|
|
@asyn.asyncf
|
|
async def get_directory(self, source_dir, dest, as_root=False):
|
|
""" Pull a directory from the device, after compressing dir """
|
|
# Create all file names
|
|
tar_file_name = source_dir.lstrip(self.path.sep).replace(self.path.sep, '.')
|
|
# Host location of dir
|
|
outdir = os.path.join(dest, tar_file_name)
|
|
# Host location of archive
|
|
tar_file_name = '{}.tar'.format(tar_file_name)
|
|
tmpfile = os.path.join(dest, tar_file_name)
|
|
|
|
# If root is required, use tmp location for tar creation.
|
|
tar_file_cm = self._xfer_cache_path if as_root else nullcontext
|
|
|
|
# Does the folder exist?
|
|
await self.execute.asyn('ls -la {}'.format(quote(source_dir)), as_root=as_root)
|
|
|
|
async with tar_file_cm(tar_file_name) as tar_file_name:
|
|
# Try compressing the folder
|
|
try:
|
|
await self.execute.asyn('{} tar -cvf {} {}'.format(
|
|
quote(self.busybox), quote(tar_file_name), quote(source_dir)
|
|
), as_root=as_root)
|
|
except TargetStableError:
|
|
self.logger.debug('Failed to run tar command on target! ' \
|
|
'Not pulling directory {}'.format(source_dir))
|
|
# Pull the file
|
|
if not os.path.exists(dest):
|
|
os.mkdir(dest)
|
|
await self.pull.asyn(tar_file_name, tmpfile)
|
|
# Decompress
|
|
with tarfile.open(tmpfile, 'r') as f:
|
|
safe_extract(f, outdir)
|
|
os.remove(tmpfile)
|
|
|
|
# execution
|
|
|
|
def _prepare_cmd(self, command, force_locale):
|
|
# Force the locale if necessary for more predictable output
|
|
if force_locale:
|
|
# Use an explicit export so that the command is allowed to be any
|
|
# shell statement, rather than just a command invocation
|
|
command = 'export LC_ALL={} && {}'.format(quote(force_locale), command)
|
|
|
|
# Ensure to use deployed command when availables
|
|
if self.executables_directory:
|
|
command = "export PATH={}:$PATH && {}".format(quote(self.executables_directory), command)
|
|
|
|
return command
|
|
|
|
class _BrokenConnection(Exception):
|
|
pass
|
|
|
|
@asyn.asyncf
|
|
@call_conn
|
|
async def _execute_async(self, *args, **kwargs):
|
|
execute = functools.partial(
|
|
self._execute,
|
|
*args, **kwargs
|
|
)
|
|
pool = self._async_pool
|
|
|
|
if pool is None:
|
|
return execute()
|
|
else:
|
|
|
|
def thread_f():
|
|
# If we cannot successfully connect from the thread, it might
|
|
# mean that something external opened a connection on the
|
|
# target, so we just revert to the blocking path.
|
|
try:
|
|
self.conn
|
|
except Exception:
|
|
raise self._BrokenConnection
|
|
else:
|
|
return execute()
|
|
|
|
loop = asyncio.get_running_loop()
|
|
try:
|
|
return await loop.run_in_executor(pool, thread_f)
|
|
except self._BrokenConnection:
|
|
return execute()
|
|
|
|
@call_conn
|
|
def _execute(self, command, timeout=None, check_exit_code=True,
|
|
as_root=False, strip_colors=True, will_succeed=False,
|
|
force_locale='C'):
|
|
|
|
command = self._prepare_cmd(command, force_locale)
|
|
return self.conn.execute(command, timeout=timeout,
|
|
check_exit_code=check_exit_code, as_root=as_root,
|
|
strip_colors=strip_colors, will_succeed=will_succeed)
|
|
|
|
execute = asyn._AsyncPolymorphicFunction(
|
|
asyn=_execute_async.asyn,
|
|
blocking=_execute,
|
|
)
|
|
|
|
@call_conn
|
|
def background(self, command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, as_root=False,
|
|
force_locale='C', timeout=None):
|
|
conn = self.conn
|
|
command = self._prepare_cmd(command, force_locale)
|
|
bg_cmd = self.conn.background(command, stdout, stderr, as_root)
|
|
if timeout is not None:
|
|
timer = threading.Timer(timeout, function=bg_cmd.cancel)
|
|
timer.daemon = True
|
|
timer.start()
|
|
return bg_cmd
|
|
|
|
def invoke(self, binary, args=None, in_directory=None, on_cpus=None,
|
|
redirect_stderr=False, as_root=False, timeout=30):
|
|
"""
|
|
Executes the specified binary under the specified conditions.
|
|
|
|
:binary: binary to execute. Must be present and executable on the device.
|
|
:args: arguments to be passed to the binary. The can be either a list or
|
|
a string.
|
|
:in_directory: execute the binary in the specified directory. This must
|
|
be an absolute path.
|
|
:on_cpus: taskset the binary to these CPUs. This may be a single ``int`` (in which
|
|
case, it will be interpreted as the mask), a list of ``ints``, in which
|
|
case this will be interpreted as the list of cpus, or string, which
|
|
will be interpreted as a comma-separated list of cpu ranges, e.g.
|
|
``"0,4-7"``.
|
|
:as_root: Specify whether the command should be run as root
|
|
:timeout: If the invocation does not terminate within this number of seconds,
|
|
a ``TimeoutError`` exception will be raised. Set to ``None`` if the
|
|
invocation should not timeout.
|
|
|
|
:returns: output of command.
|
|
"""
|
|
command = binary
|
|
if args:
|
|
if isiterable(args):
|
|
args = ' '.join(args)
|
|
command = '{} {}'.format(command, args)
|
|
if on_cpus:
|
|
on_cpus = bitmask(on_cpus)
|
|
command = '{} taskset 0x{:x} {}'.format(quote(self.busybox), on_cpus, command)
|
|
if in_directory:
|
|
command = 'cd {} && {}'.format(quote(in_directory), command)
|
|
if redirect_stderr:
|
|
command = '{} 2>&1'.format(command)
|
|
return self.execute(command, as_root=as_root, timeout=timeout)
|
|
|
|
def background_invoke(self, binary, args=None, in_directory=None,
|
|
on_cpus=None, as_root=False):
|
|
"""
|
|
Executes the specified binary as a background task under the
|
|
specified conditions.
|
|
|
|
:binary: binary to execute. Must be present and executable on the device.
|
|
:args: arguments to be passed to the binary. The can be either a list or
|
|
a string.
|
|
:in_directory: execute the binary in the specified directory. This must
|
|
be an absolute path.
|
|
:on_cpus: taskset the binary to these CPUs. This may be a single ``int`` (in which
|
|
case, it will be interpreted as the mask), a list of ``ints``, in which
|
|
case this will be interpreted as the list of cpus, or string, which
|
|
will be interpreted as a comma-separated list of cpu ranges, e.g.
|
|
``"0,4-7"``.
|
|
:as_root: Specify whether the command should be run as root
|
|
|
|
:returns: the subprocess instance handling that command
|
|
"""
|
|
command = binary
|
|
if args:
|
|
if isiterable(args):
|
|
args = ' '.join(args)
|
|
command = '{} {}'.format(command, args)
|
|
if on_cpus:
|
|
on_cpus = bitmask(on_cpus)
|
|
command = '{} taskset 0x{:x} {}'.format(quote(self.busybox), on_cpus, command)
|
|
if in_directory:
|
|
command = 'cd {} && {}'.format(quote(in_directory), command)
|
|
return self.background(command, as_root=as_root)
|
|
|
|
@asyn.asyncf
|
|
async def kick_off(self, command, as_root=None):
|
|
"""
|
|
Like execute() but returns immediately. Unlike background(), it will
|
|
not return any handle to the command being run.
|
|
"""
|
|
cmd = 'cd {wd} && {busybox} sh -c {cmd} >/dev/null 2>&1'.format(
|
|
wd=quote(self.working_directory),
|
|
busybox=quote(self.busybox),
|
|
cmd=quote(command)
|
|
)
|
|
self.background(cmd, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, as_root=as_root)
|
|
|
|
|
|
# sysfs interaction
|
|
|
|
@asyn.asyncf
|
|
async def read_value(self, path, kind=None):
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=path, mode='r')
|
|
)
|
|
output = await self.execute.asyn('cat {}'.format(quote(path)), as_root=self.needs_su) # pylint: disable=E1103
|
|
output = output.strip()
|
|
if kind:
|
|
return kind(output)
|
|
else:
|
|
return output
|
|
|
|
@asyn.asyncf
|
|
async def read_int(self, path):
|
|
return await self.read_value.asyn(path, kind=integer)
|
|
|
|
@asyn.asyncf
|
|
async def read_bool(self, path):
|
|
return await self.read_value.asyn(path, kind=boolean)
|
|
|
|
@asyn.asynccontextmanager
|
|
async def revertable_write_value(self, path, value, verify=True, as_root=True):
|
|
orig_value = self.read_value(path)
|
|
try:
|
|
await self.write_value.asyn(path, value, verify=verify, as_root=as_root)
|
|
yield
|
|
finally:
|
|
await self.write_value.asyn(path, orig_value, verify=verify, as_root=as_root)
|
|
|
|
def batch_revertable_write_value(self, kwargs_list):
|
|
return batch_contextmanager(self.revertable_write_value, kwargs_list)
|
|
|
|
@asyn.asyncf
|
|
async def write_value(self, path, value, verify=True, as_root=True):
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=path, mode='w')
|
|
)
|
|
value = str(value)
|
|
|
|
if verify:
|
|
# Check in a loop for a while since updates to sysfs files can take
|
|
# some time to be observed, typically when a write triggers a
|
|
# lengthy kernel-side request, and the read is based on some piece
|
|
# of state that may take some time to be updated by the write
|
|
# request, such as hotplugging a CPU.
|
|
cmd = '''
|
|
orig=$(cat {path} 2>/dev/null || printf "")
|
|
printf "%s" {value} > {path} || exit 10
|
|
if [ {value} != "$orig" ]; then
|
|
trials=0
|
|
while [ "$(cat {path} 2>/dev/null)" != {value} ]; do
|
|
if [ $trials -ge 10 ]; then
|
|
cat {path}
|
|
exit 11
|
|
fi
|
|
sleep 0.01
|
|
trials=$((trials + 1))
|
|
done
|
|
fi
|
|
'''
|
|
else:
|
|
cmd = '{busybox} printf "%s" {value} > {path}'
|
|
cmd = cmd.format(busybox=quote(self.busybox), path=quote(path), value=quote(value))
|
|
|
|
try:
|
|
await self.execute.asyn(cmd, check_exit_code=True, as_root=as_root)
|
|
except TargetCalledProcessError as e:
|
|
if e.returncode == 10:
|
|
raise TargetStableError('Could not write "{value}" to {path}: {e.output}'.format(
|
|
value=value, path=path, e=e))
|
|
elif verify and e.returncode == 11:
|
|
out = e.output
|
|
message = 'Could not set the value of {} to "{}" (read "{}")'.format(path, value, out)
|
|
raise TargetStableError(message)
|
|
else:
|
|
raise
|
|
|
|
@contextmanager
|
|
def make_temp(self, is_directory=True, directory='', prefix='devlib-test'):
|
|
"""
|
|
Creates temporary file/folder on target and deletes it once it's done.
|
|
|
|
:param is_directory: Specifies if temporary object is a directory, defaults to True.
|
|
:type is_directory: bool, optional
|
|
|
|
:param directory: Temp object will be created under this directory,
|
|
defaults to :attr:`Target.working_directory`.
|
|
:type directory: str, optional
|
|
|
|
:param prefix: Prefix of temp object's name, defaults to 'devlib-test'.
|
|
:type prefix: str, optional
|
|
|
|
:yield: Full path of temp object.
|
|
:rtype: str
|
|
"""
|
|
|
|
directory = directory or self.working_directory
|
|
temp_obj = None
|
|
try:
|
|
cmd = f'mktemp -p {directory} {prefix}-XXXXXX'
|
|
if is_directory:
|
|
cmd += ' -d'
|
|
|
|
temp_obj = self.execute(cmd).strip()
|
|
yield temp_obj
|
|
finally:
|
|
if temp_obj is not None:
|
|
self.remove(temp_obj)
|
|
|
|
def reset(self):
|
|
try:
|
|
self.execute('reboot', as_root=self.needs_su, timeout=2)
|
|
except (TargetError, subprocess.CalledProcessError):
|
|
# on some targets "reboot" doesn't return gracefully
|
|
pass
|
|
self.conn.connected_as_root = None
|
|
|
|
@call_conn
|
|
def check_responsive(self, explode=True):
|
|
try:
|
|
self.conn.execute('ls /', timeout=5)
|
|
return True
|
|
except (DevlibTransientError, subprocess.CalledProcessError):
|
|
if explode:
|
|
raise TargetNotRespondingError('Target {} is not responding'.format(self.conn.name))
|
|
return False
|
|
|
|
# process management
|
|
|
|
def kill(self, pid, signal=None, as_root=False):
|
|
signal_string = '-s {}'.format(signal) if signal else ''
|
|
self.execute('{} kill {} {}'.format(self.busybox, signal_string, pid), as_root=as_root)
|
|
|
|
def killall(self, process_name, signal=None, as_root=False):
|
|
for pid in self.get_pids_of(process_name):
|
|
try:
|
|
self.kill(pid, signal=signal, as_root=as_root)
|
|
except TargetStableError:
|
|
pass
|
|
|
|
def get_pids_of(self, process_name):
|
|
raise NotImplementedError()
|
|
|
|
def ps(self, **kwargs):
|
|
raise NotImplementedError()
|
|
|
|
# files
|
|
|
|
@asyn.asyncf
|
|
async def makedirs(self, path, as_root=False):
|
|
await self.execute.asyn('mkdir -p {}'.format(quote(path)), as_root=as_root)
|
|
|
|
@asyn.asyncf
|
|
async def file_exists(self, filepath):
|
|
command = 'if [ -e {} ]; then echo 1; else echo 0; fi'
|
|
output = await self.execute.asyn(command.format(quote(filepath)), as_root=self.is_rooted)
|
|
return boolean(output.strip())
|
|
|
|
@asyn.asyncf
|
|
async def directory_exists(self, filepath):
|
|
output = await self.execute.asyn('if [ -d {} ]; then echo 1; else echo 0; fi'.format(quote(filepath)))
|
|
# output from ssh my contain part of the expression in the buffer,
|
|
# split out everything except the last word.
|
|
return boolean(output.split()[-1]) # pylint: disable=maybe-no-member
|
|
|
|
@asyn.asyncf
|
|
async def list_file_systems(self):
|
|
output = await self.execute.asyn('mount')
|
|
fstab = []
|
|
for line in output.split('\n'):
|
|
line = line.strip()
|
|
if not line:
|
|
continue
|
|
match = FSTAB_ENTRY_REGEX.search(line)
|
|
if match:
|
|
fstab.append(FstabEntry(match.group(1), match.group(2),
|
|
match.group(3), match.group(4),
|
|
None, None))
|
|
else: # assume pre-M Android
|
|
fstab.append(FstabEntry(*line.split()))
|
|
return fstab
|
|
|
|
@asyn.asyncf
|
|
async def list_directory(self, path, as_root=False):
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=path, mode='r')
|
|
)
|
|
return await self._list_directory(path, as_root=as_root)
|
|
|
|
def _list_directory(self, path, as_root=False):
|
|
raise NotImplementedError()
|
|
|
|
def get_workpath(self, name):
|
|
return self.path.join(self.working_directory, name)
|
|
|
|
@asyn.asyncf
|
|
async def tempfile(self, prefix='', suffix=''):
|
|
name = '{prefix}_{uuid}_{suffix}'.format(
|
|
prefix=prefix,
|
|
uuid=uuid.uuid4().hex,
|
|
suffix=suffix,
|
|
)
|
|
path = self.get_workpath(name)
|
|
if (await self.file_exists.asyn(path)):
|
|
raise FileExistsError('Path already exists on the target: {}'.format(path))
|
|
else:
|
|
return path
|
|
|
|
@asyn.asyncf
|
|
async def remove(self, path, as_root=False):
|
|
await self.execute.asyn('rm -rf -- {}'.format(quote(path)), as_root=as_root)
|
|
|
|
# misc
|
|
@asyn.asyncf
|
|
async def read_sysctl(self, parameter):
|
|
"""
|
|
Returns the value of the given sysctl parameter as a string.
|
|
"""
|
|
path = self.path.join('/', 'proc', 'sys', *parameter.split('.'))
|
|
try:
|
|
return await self.read_value.asyn(path)
|
|
except FileNotFoundError as e:
|
|
raise ValueError(f'systcl parameter {parameter} was not found: {e}')
|
|
|
|
def core_cpus(self, core):
|
|
return [i for i, c in enumerate(self.core_names) if c == core]
|
|
|
|
@asyn.asyncf
|
|
async def list_online_cpus(self, core=None):
|
|
path = self.path.join('/sys/devices/system/cpu/online')
|
|
output = await self.read_value.asyn(path)
|
|
all_online = ranges_to_list(output)
|
|
if core:
|
|
cpus = self.core_cpus(core)
|
|
if not cpus:
|
|
raise ValueError(core)
|
|
return [o for o in all_online if o in cpus]
|
|
else:
|
|
return all_online
|
|
|
|
@asyn.asyncf
|
|
async def list_offline_cpus(self):
|
|
online = await self.list_online_cpus.asyn()
|
|
return [c for c in range(self.number_of_cpus)
|
|
if c not in online]
|
|
|
|
@asyn.asyncf
|
|
async def getenv(self, variable):
|
|
var = await self.execute.asyn('printf "%s" ${}'.format(variable))
|
|
return var.rstrip('\r\n')
|
|
|
|
def capture_screen(self, filepath):
|
|
raise NotImplementedError()
|
|
|
|
def install(self, filepath, timeout=None, with_name=None):
|
|
raise NotImplementedError()
|
|
|
|
def uninstall(self, name):
|
|
raise NotImplementedError()
|
|
|
|
@asyn.asyncf
|
|
async def get_installed(self, name, search_system_binaries=True):
|
|
# Check user installed binaries first
|
|
if self.file_exists(self.executables_directory):
|
|
if name in (await self.list_directory.asyn(self.executables_directory)):
|
|
return self.path.join(self.executables_directory, name)
|
|
# Fall back to binaries in PATH
|
|
if search_system_binaries:
|
|
PATH = await self.getenv.asyn('PATH')
|
|
for path in PATH.split(self.path.pathsep):
|
|
try:
|
|
if name in (await self.list_directory.asyn(path)):
|
|
return self.path.join(path, name)
|
|
except TargetStableError:
|
|
pass # directory does not exist or no executable permissions
|
|
|
|
which = get_installed
|
|
|
|
@asyn.asyncf
|
|
async def install_if_needed(self, host_path, search_system_binaries=True, timeout=None):
|
|
|
|
binary_path = await self.get_installed.asyn(os.path.split(host_path)[1],
|
|
search_system_binaries=search_system_binaries)
|
|
if not binary_path:
|
|
binary_path = await self.install.asyn(host_path, timeout=timeout)
|
|
return binary_path
|
|
|
|
@asyn.asyncf
|
|
async def is_installed(self, name):
|
|
return bool(await self.get_installed.asyn(name))
|
|
|
|
def bin(self, name):
|
|
return self._installed_binaries.get(name, name)
|
|
|
|
def has(self, modname):
|
|
modname = identifier(modname)
|
|
try:
|
|
self._get_module(modname, log=False)
|
|
except Exception:
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
@asyn.asyncf
|
|
async def lsmod(self):
|
|
lines = (await self.execute.asyn('lsmod')).splitlines()
|
|
entries = []
|
|
for line in lines[1:]: # first line is the header
|
|
if not line.strip():
|
|
continue
|
|
name, size, use_count, *remainder = line.split()
|
|
if remainder:
|
|
used_by = ''.join(remainder).split(',')
|
|
else:
|
|
used_by = []
|
|
entries.append(LsmodEntry(name, size, use_count, used_by))
|
|
return entries
|
|
|
|
@asyn.asyncf
|
|
async def insmod(self, path):
|
|
target_path = self.get_workpath(os.path.basename(path))
|
|
await self.push.asyn(path, target_path)
|
|
await self.execute.asyn('insmod {}'.format(quote(target_path)), as_root=True)
|
|
|
|
@asyn.asyncf
|
|
async def extract(self, path, dest=None):
|
|
"""
|
|
Extract the specified on-target file. The extraction method to be used
|
|
(unzip, gunzip, bunzip2, or tar) will be based on the file's extension.
|
|
If ``dest`` is specified, it must be an existing directory on target;
|
|
the extracted contents will be placed there.
|
|
|
|
Note that, depending on the archive file format (and therefore the
|
|
extraction method used), the original archive file may or may not exist
|
|
after the extraction.
|
|
|
|
The return value is the path to the extracted contents. In case of
|
|
gunzip and bunzip2, this will be path to the extracted file; for tar
|
|
and uzip, this will be the directory with the extracted file(s)
|
|
(``dest`` if it was specified otherwise, the directory that contained
|
|
the archive).
|
|
|
|
"""
|
|
for ending in ['.tar.gz', '.tar.bz', '.tar.bz2',
|
|
'.tgz', '.tbz', '.tbz2']:
|
|
if path.endswith(ending):
|
|
return await self._extract_archive(path, 'tar xf {} -C {}', dest)
|
|
|
|
ext = self.path.splitext(path)[1]
|
|
if ext in ['.bz', '.bz2']:
|
|
return await self._extract_file(path, 'bunzip2 -f {}', dest)
|
|
elif ext == '.gz':
|
|
return await self._extract_file(path, 'gunzip -f {}', dest)
|
|
elif ext == '.zip':
|
|
return await self._extract_archive(path, 'unzip {} -d {}', dest)
|
|
else:
|
|
raise ValueError('Unknown compression format: {}'.format(ext))
|
|
|
|
@asyn.asyncf
|
|
async def sleep(self, duration):
|
|
timeout = duration + 10
|
|
await self.execute.asyn('sleep {}'.format(duration), timeout=timeout)
|
|
|
|
@asyn.asyncf
|
|
async def read_tree_tar_flat(self, path, depth=1, check_exit_code=True,
|
|
decode_unicode=True, strip_null_chars=True):
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=path, mode='r')
|
|
)
|
|
command = 'read_tree_tgz_b64 {} {} {}'.format(quote(path), depth,
|
|
quote(self.working_directory))
|
|
output = await self._execute_util.asyn(command, as_root=self.is_rooted,
|
|
check_exit_code=check_exit_code)
|
|
|
|
result = {}
|
|
|
|
# Unpack the archive in memory
|
|
tar_gz = base64.b64decode(output)
|
|
tar_gz_bytes = io.BytesIO(tar_gz)
|
|
tar_buf = gzip.GzipFile(fileobj=tar_gz_bytes).read()
|
|
tar_bytes = io.BytesIO(tar_buf)
|
|
with tarfile.open(fileobj=tar_bytes) as tar:
|
|
for member in tar.getmembers():
|
|
try:
|
|
content_f = tar.extractfile(member)
|
|
# ignore exotic members like sockets
|
|
except Exception:
|
|
continue
|
|
# if it is a file and not a folder
|
|
if content_f:
|
|
content = content_f.read()
|
|
if decode_unicode:
|
|
try:
|
|
content = content.decode('utf-8').strip()
|
|
if strip_null_chars:
|
|
content = content.replace('\x00', '').strip()
|
|
except UnicodeDecodeError:
|
|
content = ''
|
|
|
|
name = self.path.join(path, member.name)
|
|
result[name] = content
|
|
|
|
return result
|
|
|
|
@asyn.asyncf
|
|
async def read_tree_values_flat(self, path, depth=1, check_exit_code=True):
|
|
self.async_manager.track_access(
|
|
asyn.PathAccess(namespace='target', path=path, mode='r')
|
|
)
|
|
command = 'read_tree_values {} {}'.format(quote(path), depth)
|
|
output = await self._execute_util.asyn(command, as_root=self.is_rooted,
|
|
check_exit_code=check_exit_code)
|
|
|
|
accumulator = defaultdict(list)
|
|
for entry in output.strip().split('\n'):
|
|
if ':' not in entry:
|
|
continue
|
|
path, value = entry.strip().split(':', 1)
|
|
accumulator[path].append(value)
|
|
|
|
result = {k: '\n'.join(v).strip() for k, v in accumulator.items()}
|
|
return result
|
|
|
|
@asyn.asyncf
|
|
async def read_tree_values(self, path, depth=1, dictcls=dict,
|
|
check_exit_code=True, tar=False, decode_unicode=True,
|
|
strip_null_chars=True):
|
|
"""
|
|
Reads the content of all files under a given tree
|
|
|
|
:path: path to the tree
|
|
:depth: maximum tree depth to read
|
|
:dictcls: type of the dict used to store the results
|
|
:check_exit_code: raise an exception if the shutil command fails
|
|
:tar: fetch the entire tree using tar rather than just the value (more
|
|
robust but slower in some use-cases)
|
|
:decode_unicode: decode the content of tar-ed files as utf-8
|
|
:strip_null_chars: remove '\x00' chars from the content of utf-8
|
|
decoded files
|
|
|
|
:returns: a tree-like dict with the content of files as leafs
|
|
"""
|
|
if not tar:
|
|
value_map = await self.read_tree_values_flat.asyn(path, depth, check_exit_code)
|
|
else:
|
|
value_map = await self.read_tree_tar_flat.asyn(path, depth, check_exit_code,
|
|
decode_unicode,
|
|
strip_null_chars)
|
|
return _build_path_tree(value_map, path, self.path.sep, dictcls)
|
|
|
|
def install_module(self, mod, **params):
|
|
mod = get_module(mod)
|
|
if mod.stage == 'early':
|
|
raise TargetStableError(
|
|
f'Module "{mod.name}" cannot be installed after device setup has already occoured'
|
|
)
|
|
else:
|
|
return self._install_module(mod, params)
|
|
|
|
# internal methods
|
|
|
|
@asyn.asyncf
|
|
async def _setup_shutils(self):
|
|
shutils_ifile = os.path.join(PACKAGE_BIN_DIRECTORY, 'scripts', 'shutils.in')
|
|
with open(shutils_ifile) as fh:
|
|
lines = fh.readlines()
|
|
with tempfile.TemporaryDirectory() as folder:
|
|
shutils_ofile = os.path.join(folder, 'shutils')
|
|
with open(shutils_ofile, 'w') as ofile:
|
|
for line in lines:
|
|
line = line.replace("__DEVLIB_BUSYBOX__", self.busybox)
|
|
ofile.write(line)
|
|
self._shutils = await self.install.asyn(shutils_ofile)
|
|
|
|
@asyn.asyncf
|
|
@call_conn
|
|
async def _execute_util(self, command, timeout=None, check_exit_code=True, as_root=False):
|
|
command = '{} sh {} {}'.format(quote(self.busybox), quote(self.shutils), command)
|
|
return await self.execute.asyn(
|
|
command,
|
|
timeout=timeout,
|
|
check_exit_code=check_exit_code,
|
|
as_root=as_root
|
|
)
|
|
|
|
async def _extract_archive(self, path, cmd, dest=None):
|
|
cmd = '{} ' + cmd # busybox
|
|
if dest:
|
|
extracted = dest
|
|
else:
|
|
extracted = self.path.dirname(path)
|
|
cmdtext = cmd.format(quote(self.busybox), quote(path), quote(extracted))
|
|
await self.execute.asyn(cmdtext)
|
|
return extracted
|
|
|
|
async def _extract_file(self, path, cmd, dest=None):
|
|
cmd = '{} ' + cmd # busybox
|
|
cmdtext = cmd.format(quote(self.busybox), quote(path))
|
|
await self.execute.asyn(cmdtext)
|
|
extracted = self.path.splitext(path)[0]
|
|
if dest:
|
|
await self.execute.asyn('mv -f {} {}'.format(quote(extracted), quote(dest)))
|
|
if dest.endswith('/'):
|
|
extracted = self.path.join(dest, self.path.basename(extracted))
|
|
else:
|
|
extracted = dest
|
|
return extracted
|
|
|
|
def _install_module(self, mod, params, log=True):
|
|
mod = get_module(mod)
|
|
name = mod.name
|
|
if params is None or self._modules.get(name, {}) is None:
|
|
raise TargetStableError(f'Could not load module "{name}" as it has been explicilty disabled')
|
|
else:
|
|
try:
|
|
return mod.install(self, **params)
|
|
except Exception as e:
|
|
if log:
|
|
self.logger.error(f'Module "{name}" failed to install on target: {e}')
|
|
raise
|
|
|
|
@property
|
|
def modules(self):
|
|
return sorted(self._modules.keys())
|
|
|
|
def _update_modules(self, stage):
|
|
to_install = [
|
|
(mod, params)
|
|
for mod, params in (
|
|
(get_module(name), params)
|
|
for name, params in self._modules.items()
|
|
)
|
|
if mod.stage == stage
|
|
]
|
|
for mod, params in to_install:
|
|
try:
|
|
self._install_module(mod, params)
|
|
except Exception as e:
|
|
mod_name = mod.name
|
|
self.logger.warning(f'Module {mod.name} is not supported by the target: {e}')
|
|
|
|
def _get_module(self, modname, log=True):
|
|
try:
|
|
return self._installed_modules[modname]
|
|
except KeyError:
|
|
params = {}
|
|
try:
|
|
mod = get_module(modname)
|
|
# We might try to access e.g. "boot" attribute, which is ambiguous
|
|
# since there are multiple modules with the "boot" kind. In that
|
|
# case, we look into the list of modules enabled by the user and
|
|
# get the first "boot" module we find.
|
|
except ValueError:
|
|
for _mod, _params in self._modules.items():
|
|
try:
|
|
_mod = get_module(_mod)
|
|
except ValueError:
|
|
pass
|
|
else:
|
|
if _mod.attr_name == modname:
|
|
mod = _mod
|
|
params = _params
|
|
break
|
|
else:
|
|
raise AttributeError(
|
|
f"'{self.__class__.__name__}' object has no attribute '{modname}'"
|
|
)
|
|
else:
|
|
params = self._modules.get(mod.name, {})
|
|
|
|
self._install_module(mod, params, log=log)
|
|
return self.__getattr__(modname)
|
|
|
|
def __getattr__(self, attr):
|
|
# When unpickled, objects will have an empty dict so fail early
|
|
if attr.startswith('__') and attr.endswith('__'):
|
|
raise AttributeError(attr)
|
|
|
|
try:
|
|
return self._get_module(attr)
|
|
except Exception as e:
|
|
# Raising AttributeError is important otherwise hasattr() will not
|
|
# work as expected
|
|
raise AttributeError(str(e))
|
|
|
|
def _resolve_paths(self):
|
|
raise NotImplementedError()
|
|
|
|
@asyn.asyncf
|
|
async def is_network_connected(self):
|
|
self.logger.debug('Checking for internet connectivity...')
|
|
|
|
timeout_s = 5
|
|
# It would be nice to use busybox for this, but that means we'd need
|
|
# root (ping is usually setuid so it can open raw sockets to send ICMP)
|
|
command = 'ping -q -c 1 -w {} {} 2>&1'.format(timeout_s,
|
|
quote(GOOGLE_DNS_SERVER_ADDRESS))
|
|
|
|
# We'll use our own retrying mechanism (rather than just using ping's -c
|
|
# to send multiple packets) so that we don't slow things down in the
|
|
# 'good' case where the first packet gets echoed really quickly.
|
|
attempts = 5
|
|
for _ in range(attempts):
|
|
try:
|
|
await self.execute.asyn(command)
|
|
return True
|
|
except TargetStableError as e:
|
|
err = str(e).lower()
|
|
if '100% packet loss' in err:
|
|
# We sent a packet but got no response.
|
|
# Try again - we don't want this to fail just because of a
|
|
# transient drop in connection quality.
|
|
self.logger.debug('No ping response from {} after {}s'
|
|
.format(GOOGLE_DNS_SERVER_ADDRESS, timeout_s))
|
|
continue
|
|
elif 'network is unreachable' in err:
|
|
# No internet connection at all, we can fail straight away
|
|
self.logger.debug('Network unreachable')
|
|
return False
|
|
else:
|
|
# Something else went wrong, we don't know what, raise an
|
|
# error.
|
|
raise
|
|
|
|
self.logger.debug('Failed to ping {} after {} attempts'.format(
|
|
GOOGLE_DNS_SERVER_ADDRESS, attempts))
|
|
return False
|
|
|
|
|
|
class LinuxTarget(Target):
|
|
|
|
path = posixpath
|
|
os = 'linux'
|
|
|
|
@property
|
|
@memoized
|
|
def abi(self):
|
|
value = self.execute('uname -m').strip()
|
|
for abi, architectures in ABI_MAP.items():
|
|
if value in architectures:
|
|
result = abi
|
|
break
|
|
else:
|
|
result = value
|
|
return result
|
|
|
|
@property
|
|
@memoized
|
|
def os_version(self):
|
|
os_version = {}
|
|
command = 'ls /etc/*-release /etc*-version /etc/*_release /etc/*_version 2>/dev/null'
|
|
version_files = self.execute(command, check_exit_code=False).strip().split()
|
|
for vf in version_files:
|
|
name = self.path.basename(vf)
|
|
output = self.read_value(vf)
|
|
os_version[name] = convert_new_lines(output.strip()).replace('\n', ' ')
|
|
return os_version
|
|
|
|
@property
|
|
@memoized
|
|
def system_id(self):
|
|
return self._execute_util('get_linux_system_id').strip()
|
|
|
|
def __init__(self,
|
|
connection_settings=None,
|
|
platform=None,
|
|
working_directory=None,
|
|
executables_directory=None,
|
|
connect=True,
|
|
modules=None,
|
|
load_default_modules=True,
|
|
shell_prompt=DEFAULT_SHELL_PROMPT,
|
|
conn_cls=SshConnection,
|
|
is_container=False,
|
|
max_async=50,
|
|
):
|
|
super(LinuxTarget, self).__init__(connection_settings=connection_settings,
|
|
platform=platform,
|
|
working_directory=working_directory,
|
|
executables_directory=executables_directory,
|
|
connect=connect,
|
|
modules=modules,
|
|
load_default_modules=load_default_modules,
|
|
shell_prompt=shell_prompt,
|
|
conn_cls=conn_cls,
|
|
is_container=is_container,
|
|
max_async=max_async)
|
|
|
|
def wait_boot_complete(self, timeout=10):
|
|
pass
|
|
|
|
@asyn.asyncf
|
|
async def get_pids_of(self, process_name):
|
|
"""Returns a list of PIDs of all processes with the specified name."""
|
|
# result should be a column of PIDs with the first row as "PID" header
|
|
result = await self.execute.asyn('ps -C {} -o pid'.format(quote(process_name)), # NOQA
|
|
check_exit_code=False)
|
|
result = result.strip().split()
|
|
if len(result) >= 2: # at least one row besides the header
|
|
return list(map(int, result[1:]))
|
|
else:
|
|
return []
|
|
|
|
@asyn.asyncf
|
|
async def ps(self, threads=False, **kwargs):
|
|
ps_flags = '-eo'
|
|
if threads:
|
|
ps_flags = '-eLo'
|
|
command = 'ps {} user,pid,tid,ppid,vsize,rss,wchan,pcpu,state,fname'.format(ps_flags)
|
|
|
|
out = await self.execute.asyn(command)
|
|
|
|
result = []
|
|
lines = convert_new_lines(out).splitlines()
|
|
# Skip header
|
|
for line in lines[1:]:
|
|
parts = re.split(r'\s+', line, maxsplit=9)
|
|
if parts:
|
|
result.append(PsEntry(*(parts[0:1] + list(map(int, parts[1:6])) + parts[6:])))
|
|
|
|
if not kwargs:
|
|
return result
|
|
else:
|
|
filtered_result = []
|
|
for entry in result:
|
|
if all(getattr(entry, k) == v for k, v in kwargs.items()):
|
|
filtered_result.append(entry)
|
|
return filtered_result
|
|
|
|
async def _list_directory(self, path, as_root=False):
|
|
contents = await self.execute.asyn('ls -1 {}'.format(quote(path)), as_root=as_root)
|
|
return [x.strip() for x in contents.split('\n') if x.strip()]
|
|
|
|
@asyn.asyncf
|
|
async def install(self, filepath, timeout=None, with_name=None): # pylint: disable=W0221
|
|
destpath = self.path.join(self.executables_directory,
|
|
with_name and with_name or self.path.basename(filepath))
|
|
await self.push.asyn(filepath, destpath, timeout=timeout)
|
|
await self.execute.asyn('chmod a+x {}'.format(quote(destpath)), timeout=timeout)
|
|
self._installed_binaries[self.path.basename(destpath)] = destpath
|
|
return destpath
|
|
|
|
@asyn.asyncf
|
|
async def uninstall(self, name):
|
|
path = self.path.join(self.executables_directory, name)
|
|
await self.remove.asyn(path)
|
|
|
|
@asyn.asyncf
|
|
async def capture_screen(self, filepath):
|
|
if not (await self.is_installed.asyn('scrot')):
|
|
self.logger.debug('Could not take screenshot as scrot is not installed.')
|
|
return
|
|
try:
|
|
|
|
tmpfile = await self.tempfile.asyn()
|
|
cmd = 'DISPLAY=:0.0 scrot {} && {} date -u -Iseconds'
|
|
ts = (await self.execute.asyn(cmd.format(quote(tmpfile), quote(self.busybox)))).strip()
|
|
filepath = filepath.format(ts=ts)
|
|
await self.pull.asyn(tmpfile, filepath)
|
|
await self.remove.asyn(tmpfile)
|
|
except TargetStableError as e:
|
|
if "Can't open X dispay." not in e.message:
|
|
raise e
|
|
message = e.message.split('OUTPUT:', 1)[1].strip() # pylint: disable=no-member
|
|
self.logger.debug('Could not take screenshot: {}'.format(message))
|
|
|
|
def _resolve_paths(self):
|
|
if self.working_directory is None:
|
|
self.working_directory = self.path.join(self.execute("pwd").strip(), 'devlib-target')
|
|
self._file_transfer_cache = self.path.join(self.working_directory, '.file-cache')
|
|
if self.executables_directory is None:
|
|
self.executables_directory = self.path.join(self.working_directory, 'bin')
|
|
|
|
|
|
class AndroidTarget(Target):
|
|
|
|
path = posixpath
|
|
os = 'android'
|
|
ls_command = ''
|
|
|
|
@property
|
|
@memoized
|
|
def abi(self):
|
|
return self.getprop()['ro.product.cpu.abi'].split('-')[0]
|
|
|
|
@property
|
|
@memoized
|
|
def supported_abi(self):
|
|
props = self.getprop()
|
|
result = [props['ro.product.cpu.abi']]
|
|
if 'ro.product.cpu.abi2' in props:
|
|
result.append(props['ro.product.cpu.abi2'])
|
|
if 'ro.product.cpu.abilist' in props:
|
|
for abi in props['ro.product.cpu.abilist'].split(','):
|
|
if abi not in result:
|
|
result.append(abi)
|
|
|
|
mapped_result = []
|
|
for supported_abi in result:
|
|
for abi, architectures in ABI_MAP.items():
|
|
found = False
|
|
if supported_abi in architectures and abi not in mapped_result:
|
|
mapped_result.append(abi)
|
|
found = True
|
|
break
|
|
if not found and supported_abi not in mapped_result:
|
|
mapped_result.append(supported_abi)
|
|
return mapped_result
|
|
|
|
@property
|
|
@memoized
|
|
def os_version(self):
|
|
os_version = {}
|
|
for k, v in self.getprop().iteritems():
|
|
if k.startswith('ro.build.version'):
|
|
part = k.split('.')[-1]
|
|
os_version[part] = v
|
|
return os_version
|
|
|
|
@property
|
|
def adb_name(self):
|
|
return getattr(self.conn, 'device', None)
|
|
|
|
@property
|
|
def adb_server(self):
|
|
return getattr(self.conn, 'adb_server', None)
|
|
|
|
@property
|
|
@memoized
|
|
def android_id(self):
|
|
"""
|
|
Get the device's ANDROID_ID. Which is
|
|
|
|
"A 64-bit number (as a hex string) that is randomly generated when the user
|
|
first sets up the device and should remain constant for the lifetime of the
|
|
user's device."
|
|
|
|
.. note:: This will get reset on userdata erasure.
|
|
|
|
"""
|
|
output = self.execute('content query --uri content://settings/secure --projection value --where "name=\'android_id\'"').strip()
|
|
return output.split('value=')[-1]
|
|
|
|
@property
|
|
@memoized
|
|
def system_id(self):
|
|
return self._execute_util('get_android_system_id').strip()
|
|
|
|
@property
|
|
@memoized
|
|
def external_storage(self):
|
|
return self.execute('echo $EXTERNAL_STORAGE').strip()
|
|
|
|
@property
|
|
@memoized
|
|
def external_storage_app_dir(self):
|
|
return self.path.join(self.external_storage, 'Android', 'data')
|
|
|
|
@property
|
|
@memoized
|
|
def screen_resolution(self):
|
|
output = self.execute('dumpsys window displays')
|
|
match = ANDROID_SCREEN_RESOLUTION_REGEX.search(output)
|
|
if match:
|
|
return (int(match.group('width')),
|
|
int(match.group('height')))
|
|
else:
|
|
return (0, 0)
|
|
|
|
def __init__(self,
|
|
connection_settings=None,
|
|
platform=None,
|
|
working_directory=None,
|
|
executables_directory=None,
|
|
connect=True,
|
|
modules=None,
|
|
load_default_modules=True,
|
|
shell_prompt=DEFAULT_SHELL_PROMPT,
|
|
conn_cls=AdbConnection,
|
|
package_data_directory="/data/data",
|
|
is_container=False,
|
|
max_async=50,
|
|
):
|
|
super(AndroidTarget, self).__init__(connection_settings=connection_settings,
|
|
platform=platform,
|
|
working_directory=working_directory,
|
|
executables_directory=executables_directory,
|
|
connect=connect,
|
|
modules=modules,
|
|
load_default_modules=load_default_modules,
|
|
shell_prompt=shell_prompt,
|
|
conn_cls=conn_cls,
|
|
is_container=is_container,
|
|
max_async=max_async)
|
|
self.package_data_directory = package_data_directory
|
|
self._init_logcat_lock()
|
|
|
|
def _init_logcat_lock(self):
|
|
self.clear_logcat_lock = threading.Lock()
|
|
|
|
def __getstate__(self):
|
|
dct = super().__getstate__()
|
|
return {
|
|
k: v
|
|
for k, v in dct.items()
|
|
if k not in ('clear_logcat_lock',)
|
|
}
|
|
|
|
def __setstate__(self, dct):
|
|
super().__setstate__(dct)
|
|
self._init_logcat_lock()
|
|
|
|
@asyn.asyncf
|
|
async def reset(self, fastboot=False): # pylint: disable=arguments-differ
|
|
try:
|
|
await self.execute.asyn('reboot {}'.format(fastboot and 'fastboot' or ''),
|
|
as_root=self.needs_su, timeout=2)
|
|
except (DevlibTransientError, subprocess.CalledProcessError):
|
|
# on some targets "reboot" doesn't return gracefully
|
|
pass
|
|
self.conn.connected_as_root = None
|
|
|
|
@asyn.asyncf
|
|
async def wait_boot_complete(self, timeout=10):
|
|
start = time.time()
|
|
boot_completed = boolean(await self.getprop.asyn('sys.boot_completed'))
|
|
while not boot_completed and timeout >= time.time() - start:
|
|
time.sleep(5)
|
|
boot_completed = boolean(await self.getprop.asyn('sys.boot_completed'))
|
|
if not boot_completed:
|
|
# Raise a TargetStableError as this usually happens because of
|
|
# an issue with Android more than a timeout that is too small.
|
|
raise TargetStableError('Connected but Android did not fully boot.')
|
|
|
|
@asyn.asyncf
|
|
async def connect(self, timeout=30, check_boot_completed=True, max_async=None): # pylint: disable=arguments-differ
|
|
device = self.connection_settings.get('device')
|
|
await super(AndroidTarget, self).connect.asyn(
|
|
timeout=timeout,
|
|
check_boot_completed=check_boot_completed,
|
|
max_async=max_async,
|
|
)
|
|
|
|
@asyn.asyncf
|
|
async def __setup_list_directory(self):
|
|
# In at least Linaro Android 16.09 (which was their first Android 7 release) and maybe
|
|
# AOSP 7.0 as well, the ls command was changed.
|
|
# Previous versions default to a single column listing, which is nice and easy to parse.
|
|
# Newer versions default to a multi-column listing, which is not, but it does support
|
|
# a '-1' option to get into single column mode. Older versions do not support this option
|
|
# so we try the new version, and if it fails we use the old version.
|
|
self.ls_command = 'ls -1'
|
|
try:
|
|
await self.execute.asyn('ls -1 {}'.format(quote(self.working_directory)), as_root=False)
|
|
except TargetStableError:
|
|
self.ls_command = 'ls'
|
|
|
|
async def _list_directory(self, path, as_root=False):
|
|
if self.ls_command == '':
|
|
await self.__setup_list_directory.asyn()
|
|
contents = await self.execute.asyn('{} {}'.format(self.ls_command, quote(path)), as_root=as_root)
|
|
return [x.strip() for x in contents.split('\n') if x.strip()]
|
|
|
|
@asyn.asyncf
|
|
async def install(self, filepath, timeout=None, with_name=None): # pylint: disable=W0221
|
|
ext = os.path.splitext(filepath)[1].lower()
|
|
if ext == '.apk':
|
|
return await self.install_apk.asyn(filepath, timeout)
|
|
else:
|
|
return await self.install_executable.asyn(filepath, with_name, timeout)
|
|
|
|
@asyn.asyncf
|
|
async def uninstall(self, name):
|
|
if await self.package_is_installed.asyn(name):
|
|
await self.uninstall_package.asyn(name)
|
|
else:
|
|
await self.uninstall_executable.asyn(name)
|
|
|
|
@asyn.asyncf
|
|
async def get_pids_of(self, process_name):
|
|
result = []
|
|
search_term = process_name[-15:]
|
|
for entry in await self.ps.asyn():
|
|
if search_term in entry.name:
|
|
result.append(entry.pid)
|
|
return result
|
|
|
|
@asyn.asyncf
|
|
async def ps(self, threads=False, **kwargs):
|
|
maxsplit = 9 if threads else 8
|
|
command = 'ps'
|
|
if threads:
|
|
command = 'ps -AT'
|
|
|
|
lines = iter(convert_new_lines(await self.execute.asyn(command)).split('\n'))
|
|
next(lines) # header
|
|
result = []
|
|
for line in lines:
|
|
parts = line.split(None, maxsplit)
|
|
if not parts:
|
|
continue
|
|
|
|
wchan_missing = False
|
|
if len(parts) == maxsplit:
|
|
wchan_missing = True
|
|
|
|
if not threads:
|
|
# Duplicate PID into TID location.
|
|
parts.insert(2, parts[1])
|
|
|
|
if wchan_missing:
|
|
# wchan was blank; insert an empty field where it should be.
|
|
parts.insert(6, '')
|
|
result.append(PsEntry(*(parts[0:1] + list(map(int, parts[1:6])) + parts[6:])))
|
|
if not kwargs:
|
|
return result
|
|
else:
|
|
filtered_result = []
|
|
for entry in result:
|
|
if all(getattr(entry, k) == v for k, v in kwargs.items()):
|
|
filtered_result.append(entry)
|
|
return filtered_result
|
|
|
|
@asyn.asyncf
|
|
async def capture_screen(self, filepath):
|
|
on_device_file = self.path.join(self.working_directory, 'screen_capture.png')
|
|
cmd = 'screencap -p {} && {} date -u -Iseconds'
|
|
ts = (await self.execute.asyn(cmd.format(quote(on_device_file), quote(self.busybox)))).strip()
|
|
filepath = filepath.format(ts=ts)
|
|
await self.pull.asyn(on_device_file, filepath)
|
|
await self.remove.asyn(on_device_file)
|
|
|
|
# Android-specific
|
|
|
|
@asyn.asyncf
|
|
async def input_tap(self, x, y):
|
|
command = 'input tap {} {}'
|
|
await self.execute.asyn(command.format(x, y))
|
|
|
|
@asyn.asyncf
|
|
async def input_tap_pct(self, x, y):
|
|
width, height = self.screen_resolution
|
|
|
|
x = (x * width) // 100
|
|
y = (y * height) // 100
|
|
|
|
await self.input_tap.asyn(x, y)
|
|
|
|
@asyn.asyncf
|
|
async def input_swipe(self, x1, y1, x2, y2):
|
|
"""
|
|
Issue a swipe on the screen from (x1, y1) to (x2, y2)
|
|
Uses absolute screen positions
|
|
"""
|
|
command = 'input swipe {} {} {} {}'
|
|
await self.execute.asyn(command.format(x1, y1, x2, y2))
|
|
|
|
@asyn.asyncf
|
|
async def input_swipe_pct(self, x1, y1, x2, y2):
|
|
"""
|
|
Issue a swipe on the screen from (x1, y1) to (x2, y2)
|
|
Uses percent-based positions
|
|
"""
|
|
width, height = self.screen_resolution
|
|
|
|
x1 = (x1 * width) // 100
|
|
y1 = (y1 * height) // 100
|
|
x2 = (x2 * width) // 100
|
|
y2 = (y2 * height) // 100
|
|
|
|
await self.input_swipe.asyn(x1, y1, x2, y2)
|
|
|
|
@asyn.asyncf
|
|
async def swipe_to_unlock(self, direction="diagonal"):
|
|
width, height = self.screen_resolution
|
|
if direction == "diagonal":
|
|
start = 100
|
|
stop = width - start
|
|
swipe_height = height * 2 // 3
|
|
await self.input_swipe.asyn(start, swipe_height, stop, 0)
|
|
elif direction == "horizontal":
|
|
swipe_height = height * 2 // 3
|
|
start = 100
|
|
stop = width - start
|
|
await self.input_swipe.asyn(start, swipe_height, stop, swipe_height)
|
|
elif direction == "vertical":
|
|
swipe_middle = width / 2
|
|
swipe_height = height * 2 // 3
|
|
await self.input_swipe.asyn(swipe_middle, swipe_height, swipe_middle, 0)
|
|
else:
|
|
raise TargetStableError("Invalid swipe direction: {}".format(direction))
|
|
|
|
@asyn.asyncf
|
|
async def getprop(self, prop=None):
|
|
props = AndroidProperties(await self.execute.asyn('getprop'))
|
|
if prop:
|
|
return props[prop]
|
|
return props
|
|
|
|
@asyn.asyncf
|
|
async def capture_ui_hierarchy(self, filepath):
|
|
on_target_file = self.get_workpath('screen_capture.xml')
|
|
try:
|
|
await self.execute.asyn('uiautomator dump {}'.format(on_target_file))
|
|
await self.pull.asyn(on_target_file, filepath)
|
|
finally:
|
|
await self.remove.asyn(on_target_file)
|
|
|
|
parsed_xml = xml.dom.minidom.parse(filepath)
|
|
with open(filepath, 'w') as f:
|
|
f.write(parsed_xml.toprettyxml())
|
|
|
|
@asyn.asyncf
|
|
async def is_installed(self, name):
|
|
return (await super(AndroidTarget, self).is_installed.asyn(name)) or (await self.package_is_installed.asyn(name))
|
|
|
|
@asyn.asyncf
|
|
async def package_is_installed(self, package_name):
|
|
return package_name in (await self.list_packages.asyn())
|
|
|
|
@asyn.asyncf
|
|
async def list_packages(self):
|
|
output = await self.execute.asyn('pm list packages')
|
|
output = output.replace('package:', '')
|
|
return output.split()
|
|
|
|
@asyn.asyncf
|
|
async def get_package_version(self, package):
|
|
output = await self.execute.asyn('dumpsys package {}'.format(quote(package)))
|
|
for line in convert_new_lines(output).split('\n'):
|
|
if 'versionName' in line:
|
|
return line.split('=', 1)[1]
|
|
return None
|
|
|
|
@asyn.asyncf
|
|
async def get_package_info(self, package):
|
|
output = await self.execute.asyn('pm list packages -f {}'.format(quote(package)))
|
|
for entry in output.strip().split('\n'):
|
|
rest, entry_package = entry.rsplit('=', 1)
|
|
if entry_package != package:
|
|
continue
|
|
_, apk_path = rest.split(':')
|
|
return installed_package_info(apk_path, entry_package)
|
|
|
|
@asyn.asyncf
|
|
async def get_sdk_version(self):
|
|
try:
|
|
return int(await self.getprop.asyn('ro.build.version.sdk'))
|
|
except (ValueError, TypeError):
|
|
return None
|
|
|
|
@asyn.asyncf
|
|
async def install_apk(self, filepath, timeout=None, replace=False, allow_downgrade=False): # pylint: disable=W0221
|
|
ext = os.path.splitext(filepath)[1].lower()
|
|
if ext == '.apk':
|
|
flags = []
|
|
if replace:
|
|
flags.append('-r') # Replace existing APK
|
|
if allow_downgrade:
|
|
flags.append('-d') # Install the APK even if a newer version is already installed
|
|
if self.get_sdk_version() >= 23:
|
|
flags.append('-g') # Grant all runtime permissions
|
|
self.logger.debug("Replace APK = {}, ADB flags = '{}'".format(replace, ' '.join(flags)))
|
|
if isinstance(self.conn, AdbConnection):
|
|
return adb_command(self.adb_name, "install {} {}".format(' '.join(flags), quote(filepath)),
|
|
timeout=timeout, adb_server=self.adb_server)
|
|
else:
|
|
dev_path = self.get_workpath(filepath.rsplit(os.path.sep, 1)[-1])
|
|
await self.push.asyn(quote(filepath), dev_path, timeout=timeout)
|
|
result = await self.execute.asyn("pm install {} {}".format(' '.join(flags), quote(dev_path)), timeout=timeout)
|
|
await self.remove.asyn(dev_path)
|
|
return result
|
|
else:
|
|
raise TargetStableError('Can\'t install {}: unsupported format.'.format(filepath))
|
|
|
|
@asyn.asyncf
|
|
async def grant_package_permission(self, package, permission):
|
|
try:
|
|
return await self.execute.asyn('pm grant {} {}'.format(quote(package), quote(permission)))
|
|
except TargetStableError as e:
|
|
if 'is not a changeable permission type' in e.message:
|
|
pass # Ignore if unchangeable
|
|
elif 'Unknown permission' in e.message:
|
|
pass # Ignore if unknown
|
|
elif 'has not requested permission' in e.message:
|
|
pass # Ignore if not requested
|
|
elif 'Operation not allowed' in e.message:
|
|
pass # Ignore if not allowed
|
|
elif 'is managed by role' in e.message:
|
|
pass # Ignore if cannot be granted
|
|
else:
|
|
raise
|
|
|
|
@asyn.asyncf
|
|
async def refresh_files(self, file_list):
|
|
"""
|
|
Depending on the android version and root status, determine the
|
|
appropriate method of forcing a re-index of the mediaserver cache for a given
|
|
list of files.
|
|
"""
|
|
if self.is_rooted or (await self.get_sdk_version.asyn()) < 24: # MM and below
|
|
common_path = commonprefix(file_list, sep=self.path.sep)
|
|
await self.broadcast_media_mounted.asyn(common_path, self.is_rooted)
|
|
else:
|
|
for f in file_list:
|
|
await self.broadcast_media_scan_file.asyn(f)
|
|
|
|
@asyn.asyncf
|
|
async def broadcast_media_scan_file(self, filepath):
|
|
"""
|
|
Force a re-index of the mediaserver cache for the specified file.
|
|
"""
|
|
command = 'am broadcast -a android.intent.action.MEDIA_SCANNER_SCAN_FILE -d {}'
|
|
await self.execute.asyn(command.format(quote('file://' + filepath)))
|
|
|
|
@asyn.asyncf
|
|
async def broadcast_media_mounted(self, dirpath, as_root=False):
|
|
"""
|
|
Force a re-index of the mediaserver cache for the specified directory.
|
|
"""
|
|
command = 'am broadcast -a android.intent.action.MEDIA_MOUNTED -d {} '\
|
|
'-n com.android.providers.media/.MediaScannerReceiver'
|
|
await self.execute.asyn(command.format(quote('file://'+dirpath)), as_root=as_root)
|
|
|
|
@asyn.asyncf
|
|
async def install_executable(self, filepath, with_name=None, timeout=None):
|
|
self._ensure_executables_directory_is_writable()
|
|
executable_name = with_name or os.path.basename(filepath)
|
|
on_device_file = self.path.join(self.working_directory, executable_name)
|
|
on_device_executable = self.path.join(self.executables_directory, executable_name)
|
|
await self.push.asyn(filepath, on_device_file, timeout=timeout)
|
|
if on_device_file != on_device_executable:
|
|
await self.execute.asyn('cp -f -- {} {}'.format(quote(on_device_file), quote(on_device_executable)),
|
|
as_root=self.needs_su, timeout=timeout)
|
|
await self.remove.asyn(on_device_file, as_root=self.needs_su)
|
|
await self.execute.asyn("chmod 0777 {}".format(quote(on_device_executable)), as_root=self.needs_su)
|
|
self._installed_binaries[executable_name] = on_device_executable
|
|
return on_device_executable
|
|
|
|
@asyn.asyncf
|
|
async def uninstall_package(self, package):
|
|
if isinstance(self.conn, AdbConnection):
|
|
adb_command(self.adb_name, "uninstall {}".format(quote(package)), timeout=30,
|
|
adb_server=self.adb_server)
|
|
else:
|
|
await self.execute.asyn("pm uninstall {}".format(quote(package)), timeout=30)
|
|
|
|
@asyn.asyncf
|
|
async def uninstall_executable(self, executable_name):
|
|
on_device_executable = self.path.join(self.executables_directory, executable_name)
|
|
self._ensure_executables_directory_is_writable()
|
|
await self.remove.asyn(on_device_executable, as_root=self.needs_su)
|
|
|
|
@asyn.asyncf
|
|
async def dump_logcat(self, filepath, filter=None, logcat_format=None, append=False,
|
|
timeout=60): # pylint: disable=redefined-builtin
|
|
op = '>>' if append else '>'
|
|
filtstr = ' -s {}'.format(quote(filter)) if filter else ''
|
|
formatstr = ' -v {}'.format(quote(logcat_format)) if logcat_format else ''
|
|
logcat_opts = '-d' + formatstr + filtstr
|
|
if isinstance(self.conn, AdbConnection):
|
|
command = 'logcat {} {} {}'.format(logcat_opts, op, quote(filepath))
|
|
adb_command(self.adb_name, command, timeout=timeout, adb_server=self.adb_server)
|
|
else:
|
|
dev_path = self.get_workpath('logcat')
|
|
command = 'logcat {} {} {}'.format(logcat_opts, op, quote(dev_path))
|
|
await self.execute.asyn(command, timeout=timeout)
|
|
await self.pull.asyn(dev_path, filepath)
|
|
await self.remove.asyn(dev_path)
|
|
|
|
@asyn.asyncf
|
|
async def clear_logcat(self):
|
|
locked = self.clear_logcat_lock.acquire(blocking=False)
|
|
if locked:
|
|
try:
|
|
if isinstance(self.conn, AdbConnection):
|
|
adb_command(self.adb_name, 'logcat -c', timeout=30, adb_server=self.adb_server)
|
|
else:
|
|
await self.execute.asyn('logcat -c', timeout=30)
|
|
finally:
|
|
self.clear_logcat_lock.release()
|
|
|
|
def get_logcat_monitor(self, regexps=None):
|
|
return LogcatMonitor(self, regexps)
|
|
|
|
@call_conn
|
|
def wait_for_device(self, timeout=30):
|
|
self.conn.wait_for_device()
|
|
|
|
@call_conn
|
|
def reboot_bootloader(self, timeout=30):
|
|
self.conn.reboot_bootloader()
|
|
|
|
@asyn.asyncf
|
|
async def is_screen_locked(self):
|
|
screen_state = await self.execute.asyn('dumpsys window')
|
|
return 'mDreamingLockscreen=true' in screen_state
|
|
|
|
@asyn.asyncf
|
|
async def is_screen_on(self):
|
|
output = await self.execute.asyn('dumpsys power')
|
|
match = ANDROID_SCREEN_STATE_REGEX.search(output)
|
|
if match:
|
|
if 'DOZE' in match.group(1).upper():
|
|
return True
|
|
if match.group(1) == 'Dozing':
|
|
return False
|
|
if match.group(1) == 'Asleep':
|
|
return False
|
|
if match.group(1) == 'Awake':
|
|
return True
|
|
return boolean(match.group(1))
|
|
else:
|
|
raise TargetStableError('Could not establish screen state.')
|
|
|
|
@asyn.asyncf
|
|
async def ensure_screen_is_on(self, verify=True):
|
|
if not await self.is_screen_on.asyn():
|
|
self.execute('input keyevent 26')
|
|
if verify and not await self.is_screen_on.asyn():
|
|
raise TargetStableError('Display cannot be turned on.')
|
|
|
|
@asyn.asyncf
|
|
async def ensure_screen_is_on_and_stays(self, verify=True, mode=7):
|
|
await self.ensure_screen_is_on.asyn(verify=verify)
|
|
await self.set_stay_on_mode.asyn(mode)
|
|
|
|
@asyn.asyncf
|
|
async def ensure_screen_is_off(self, verify=True):
|
|
# Allow 2 attempts to help with cases of ambient display modes
|
|
# where the first attempt will switch the display fully on.
|
|
for _ in range(2):
|
|
if await self.is_screen_on.asyn():
|
|
await self.execute.asyn('input keyevent 26')
|
|
time.sleep(0.5)
|
|
if verify and await self.is_screen_on.asyn():
|
|
msg = 'Display cannot be turned off. Is always on display enabled?'
|
|
raise TargetStableError(msg)
|
|
|
|
@asyn.asyncf
|
|
async def set_auto_brightness(self, auto_brightness):
|
|
cmd = 'settings put system screen_brightness_mode {}'
|
|
await self.execute.asyn(cmd.format(int(boolean(auto_brightness))))
|
|
|
|
@asyn.asyncf
|
|
async def get_auto_brightness(self):
|
|
cmd = 'settings get system screen_brightness_mode'
|
|
return boolean((await self.execute.asyn(cmd)).strip())
|
|
|
|
@asyn.asyncf
|
|
async def set_brightness(self, value):
|
|
if not 0 <= value <= 255:
|
|
msg = 'Invalid brightness "{}"; Must be between 0 and 255'
|
|
raise ValueError(msg.format(value))
|
|
await self.set_auto_brightness.asyn(False)
|
|
cmd = 'settings put system screen_brightness {}'
|
|
await self.execute.asyn(cmd.format(int(value)))
|
|
|
|
@asyn.asyncf
|
|
async def get_brightness(self):
|
|
cmd = 'settings get system screen_brightness'
|
|
return integer((await self.execute.asyn(cmd)).strip())
|
|
|
|
@asyn.asyncf
|
|
async def set_screen_timeout(self, timeout_ms):
|
|
cmd = 'settings put system screen_off_timeout {}'
|
|
await self.execute.asyn(cmd.format(int(timeout_ms)))
|
|
|
|
@asyn.asyncf
|
|
async def get_screen_timeout(self):
|
|
cmd = 'settings get system screen_off_timeout'
|
|
return int((await self.execute.asyn(cmd)).strip())
|
|
|
|
@asyn.asyncf
|
|
async def get_airplane_mode(self):
|
|
cmd = 'settings get global airplane_mode_on'
|
|
return boolean((await self.execute.asyn(cmd)).strip())
|
|
|
|
@asyn.asyncf
|
|
async def get_stay_on_mode(self):
|
|
cmd = 'settings get global stay_on_while_plugged_in'
|
|
return int((await self.execute.asyn(cmd)).strip())
|
|
|
|
@asyn.asyncf
|
|
async def set_airplane_mode(self, mode):
|
|
root_required = await self.get_sdk_version.asyn() > 23
|
|
if root_required and not self.is_rooted:
|
|
raise TargetStableError('Root is required to toggle airplane mode on Android 7+')
|
|
mode = int(boolean(mode))
|
|
cmd = 'settings put global airplane_mode_on {}'
|
|
await self.execute.asyn(cmd.format(mode))
|
|
await self.execute.asyn('am broadcast -a android.intent.action.AIRPLANE_MODE '
|
|
'--ez state {}'.format(mode), as_root=root_required)
|
|
|
|
@asyn.asyncf
|
|
async def get_auto_rotation(self):
|
|
cmd = 'settings get system accelerometer_rotation'
|
|
return boolean((await self.execute.asyn(cmd)).strip())
|
|
|
|
@asyn.asyncf
|
|
async def set_auto_rotation(self, autorotate):
|
|
cmd = 'settings put system accelerometer_rotation {}'
|
|
await self.execute.asyn(cmd.format(int(boolean(autorotate))))
|
|
|
|
@asyn.asyncf
|
|
async def set_natural_rotation(self):
|
|
await self.set_rotation.asyn(0)
|
|
|
|
@asyn.asyncf
|
|
async def set_left_rotation(self):
|
|
await self.set_rotation.asyn(1)
|
|
|
|
@asyn.asyncf
|
|
async def set_inverted_rotation(self):
|
|
await self.set_rotation.asyn(2)
|
|
|
|
@asyn.asyncf
|
|
async def set_right_rotation(self):
|
|
await self.set_rotation.asyn(3)
|
|
|
|
@asyn.asyncf
|
|
async def get_rotation(self):
|
|
output = await self.execute.asyn('dumpsys input')
|
|
match = ANDROID_SCREEN_ROTATION_REGEX.search(output)
|
|
if match:
|
|
return int(match.group('rotation'))
|
|
else:
|
|
return None
|
|
|
|
@asyn.asyncf
|
|
async def set_rotation(self, rotation):
|
|
if not 0 <= rotation <= 3:
|
|
raise ValueError('Rotation value must be between 0 and 3')
|
|
await self.set_auto_rotation.asyn(False)
|
|
cmd = 'settings put system user_rotation {}'
|
|
await self.execute.asyn(cmd.format(rotation))
|
|
|
|
@asyn.asyncf
|
|
async def set_stay_on_never(self):
|
|
await self.set_stay_on_mode.asyn(0)
|
|
|
|
@asyn.asyncf
|
|
async def set_stay_on_while_powered(self):
|
|
await self.set_stay_on_mode.asyn(7)
|
|
|
|
@asyn.asyncf
|
|
async def set_stay_on_mode(self, mode):
|
|
if not 0 <= mode <= 7:
|
|
raise ValueError('Screen stay on mode must be between 0 and 7')
|
|
cmd = 'settings put global stay_on_while_plugged_in {}'
|
|
await self.execute.asyn(cmd.format(mode))
|
|
|
|
@asyn.asyncf
|
|
async def open_url(self, url, force_new=False):
|
|
"""
|
|
Start a view activity by specifying an URL
|
|
|
|
:param url: URL of the item to display
|
|
:type url: str
|
|
|
|
:param force_new: Force the viewing application to be relaunched
|
|
if it is already running
|
|
:type force_new: bool
|
|
"""
|
|
cmd = 'am start -a android.intent.action.VIEW -d {}'
|
|
|
|
if force_new:
|
|
cmd = cmd + ' -f {}'.format(INTENT_FLAGS['ACTIVITY_NEW_TASK'] |
|
|
INTENT_FLAGS['ACTIVITY_CLEAR_TASK'])
|
|
|
|
await self.execute.asyn(cmd.format(quote(url)))
|
|
|
|
@asyn.asyncf
|
|
async def homescreen(self):
|
|
await self.execute.asyn('am start -a android.intent.action.MAIN -c android.intent.category.HOME')
|
|
|
|
def _resolve_paths(self):
|
|
if self.working_directory is None:
|
|
self.working_directory = self.path.join(self.external_storage, 'devlib-target')
|
|
self._file_transfer_cache = self.path.join(self.working_directory, '.file-cache')
|
|
if self.executables_directory is None:
|
|
self.executables_directory = '/data/local/tmp/bin'
|
|
|
|
@asyn.asyncf
|
|
async def _ensure_executables_directory_is_writable(self):
|
|
matched = []
|
|
for entry in await self.list_file_systems.asyn():
|
|
if self.executables_directory.rstrip('/').startswith(entry.mount_point):
|
|
matched.append(entry)
|
|
if matched:
|
|
entry = sorted(matched, key=lambda x: len(x.mount_point))[-1]
|
|
if 'rw' not in entry.options:
|
|
await self.execute.asyn('mount -o rw,remount {} {}'.format(quote(entry.device),
|
|
quote(entry.mount_point)),
|
|
as_root=True)
|
|
else:
|
|
message = 'Could not find mount point for executables directory {}'
|
|
raise TargetStableError(message.format(self.executables_directory))
|
|
|
|
_charging_enabled_path = '/sys/class/power_supply/battery/charging_enabled'
|
|
|
|
@property
|
|
def charging_enabled(self):
|
|
"""
|
|
Whether drawing power to charge the battery is enabled
|
|
|
|
Not all devices have the ability to enable/disable battery charging
|
|
(e.g. because they don't have a battery). In that case,
|
|
``charging_enabled`` is None.
|
|
"""
|
|
if not self.file_exists(self._charging_enabled_path):
|
|
return None
|
|
return self.read_bool(self._charging_enabled_path)
|
|
|
|
@charging_enabled.setter
|
|
def charging_enabled(self, enabled):
|
|
"""
|
|
Enable/disable drawing power to charge the battery
|
|
|
|
Not all devices have this facility. In that case, do nothing.
|
|
"""
|
|
if not self.file_exists(self._charging_enabled_path):
|
|
return
|
|
self.write_value(self._charging_enabled_path, int(bool(enabled)))
|
|
|
|
FstabEntry = namedtuple('FstabEntry', ['device', 'mount_point', 'fs_type', 'options', 'dump_freq', 'pass_num'])
|
|
PsEntry = namedtuple('PsEntry', 'user pid tid ppid vsize rss wchan pc state name')
|
|
LsmodEntry = namedtuple('LsmodEntry', ['name', 'size', 'use_count', 'used_by'])
|
|
|
|
|
|
class Cpuinfo(object):
|
|
|
|
@property
|
|
@memoized
|
|
def architecture(self):
|
|
for section in self.sections:
|
|
if 'CPU architecture' in section:
|
|
return section['CPU architecture']
|
|
if 'architecture' in section:
|
|
return section['architecture']
|
|
|
|
@property
|
|
@memoized
|
|
def cpu_names(self):
|
|
cpu_names = []
|
|
global_name = None
|
|
for section in self.sections:
|
|
if 'processor' in section:
|
|
if 'CPU part' in section:
|
|
cpu_names.append(_get_part_name(section))
|
|
elif 'model name' in section:
|
|
cpu_names.append(_get_model_name(section))
|
|
else:
|
|
cpu_names.append(None)
|
|
elif 'CPU part' in section:
|
|
global_name = _get_part_name(section)
|
|
return [caseless_string(c or global_name) for c in cpu_names]
|
|
|
|
def __init__(self, text):
|
|
self.sections = None
|
|
self.text = None
|
|
self.parse(text)
|
|
|
|
@memoized
|
|
def get_cpu_features(self, cpuid=0):
|
|
global_features = []
|
|
for section in self.sections:
|
|
if 'processor' in section:
|
|
if int(section.get('processor')) != cpuid:
|
|
continue
|
|
if 'Features' in section:
|
|
return section.get('Features').split()
|
|
elif 'flags' in section:
|
|
return section.get('flags').split()
|
|
elif 'Features' in section:
|
|
global_features = section.get('Features').split()
|
|
elif 'flags' in section:
|
|
global_features = section.get('flags').split()
|
|
return global_features
|
|
|
|
def parse(self, text):
|
|
self.sections = []
|
|
current_section = {}
|
|
self.text = text.strip()
|
|
for line in self.text.split('\n'):
|
|
line = line.strip()
|
|
if line:
|
|
key, value = line.split(':', 1)
|
|
current_section[key.strip()] = value.strip()
|
|
else: # not line
|
|
self.sections.append(current_section)
|
|
current_section = {}
|
|
self.sections.append(current_section)
|
|
|
|
def __str__(self):
|
|
return 'CpuInfo({})'.format(self.cpu_names)
|
|
|
|
__repr__ = __str__
|
|
|
|
|
|
class KernelVersion(object):
|
|
"""
|
|
Class representing the version of a target kernel
|
|
|
|
Not expected to work for very old (pre-3.0) kernel version numbers.
|
|
|
|
:ivar release: Version number/revision string. Typical output of
|
|
``uname -r``
|
|
:type release: str
|
|
:ivar version: Extra version info (aside from ``release``) reported by
|
|
``uname``
|
|
:type version: str
|
|
:ivar version_number: Main version number (e.g. 3 for Linux 3.18)
|
|
:type version_number: int
|
|
:ivar major: Major version number (e.g. 18 for Linux 3.18)
|
|
:type major: int
|
|
:ivar minor: Minor version number for stable kernels (e.g. 9 for 4.9.9). May
|
|
be None
|
|
:type minor: int
|
|
:ivar rc: Release candidate number (e.g. 3 for Linux 4.9-rc3). May be None.
|
|
:type rc: int
|
|
:ivar commits: Number of additional commits on the branch. May be None.
|
|
:type commits: int
|
|
:ivar sha1: Kernel git revision hash, if available (otherwise None)
|
|
:type sha1: str
|
|
:ivar android_version: Android version, if available (otherwise None)
|
|
:type android_version: int
|
|
:ivar gki_abi: GKI kernel abi, if available (otherwise None)
|
|
:type gki_abi: str
|
|
|
|
:ivar parts: Tuple of version number components. Can be used for
|
|
lexicographically comparing kernel versions.
|
|
:type parts: tuple(int)
|
|
"""
|
|
def __init__(self, version_string):
|
|
if ' #' in version_string:
|
|
release, version = version_string.split(' #')
|
|
self.release = release
|
|
self.version = version
|
|
elif version_string.startswith('#'):
|
|
self.release = ''
|
|
self.version = version_string
|
|
else:
|
|
self.release = version_string
|
|
self.version = ''
|
|
|
|
self.version_number = None
|
|
self.major = None
|
|
self.minor = None
|
|
self.sha1 = None
|
|
self.rc = None
|
|
self.commits = None
|
|
self.gki_abi = None
|
|
self.android_version = None
|
|
match = KVERSION_REGEX.match(version_string)
|
|
if match:
|
|
groups = match.groupdict()
|
|
self.version_number = int(groups['version'])
|
|
self.major = int(groups['major'])
|
|
if groups['minor'] is not None:
|
|
self.minor = int(groups['minor'])
|
|
if groups['rc'] is not None:
|
|
self.rc = int(groups['rc'])
|
|
if groups['commits'] is not None:
|
|
self.commits = int(groups['commits'])
|
|
if groups['sha1'] is not None:
|
|
self.sha1 = match.group('sha1')
|
|
if groups['gki_abi'] is not None:
|
|
self.gki_abi = match.group('gki_abi')
|
|
if groups['android_version'] is not None:
|
|
self.android_version = int(match.group('android_version'))
|
|
|
|
self.parts = (self.version_number, self.major, self.minor)
|
|
|
|
def __str__(self):
|
|
return '{} {}'.format(self.release, self.version)
|
|
|
|
__repr__ = __str__
|
|
|
|
|
|
class HexInt(long):
|
|
"""
|
|
Subclass of :class:`int` that uses hexadecimal formatting by default.
|
|
"""
|
|
|
|
def __new__(cls, val=0, base=16):
|
|
super_new = super(HexInt, cls).__new__
|
|
if isinstance(val, Number):
|
|
return super_new(cls, val)
|
|
else:
|
|
return super_new(cls, val, base=base)
|
|
|
|
def __str__(self):
|
|
return hex(self).strip('L')
|
|
|
|
|
|
class KernelConfigTristate(Enum):
|
|
YES = 'y'
|
|
NO = 'n'
|
|
MODULE = 'm'
|
|
|
|
def __bool__(self):
|
|
"""
|
|
Allow using this enum to represent bool Kconfig type, although it is
|
|
technically different from tristate.
|
|
"""
|
|
return self in (self.YES, self.MODULE)
|
|
|
|
def __nonzero__(self):
|
|
"""
|
|
For Python 2.x compatibility.
|
|
"""
|
|
return self.__bool__()
|
|
|
|
@classmethod
|
|
def from_str(cls, str_):
|
|
for state in cls:
|
|
if state.value == str_:
|
|
return state
|
|
raise ValueError('No kernel config tristate value matches "{}"'.format(str_))
|
|
|
|
|
|
class TypedKernelConfig(Mapping):
|
|
"""
|
|
Mapping-like typed version of :class:`KernelConfig`.
|
|
|
|
Values are either :class:`str`, :class:`int`,
|
|
:class:`KernelConfigTristate`, or :class:`HexInt`. ``hex`` Kconfig type is
|
|
mapped to :class:`HexInt` and ``bool`` to :class:`KernelConfigTristate`.
|
|
"""
|
|
not_set_regex = re.compile(r'# (\S+) is not set')
|
|
|
|
@staticmethod
|
|
def get_config_name(name):
|
|
name = name.upper()
|
|
if not name.startswith('CONFIG_'):
|
|
name = 'CONFIG_' + name
|
|
return name
|
|
|
|
def __init__(self, mapping=None):
|
|
mapping = mapping if mapping is not None else {}
|
|
self._config = {
|
|
# Ensure we use the canonical name of the config keys for internal
|
|
# representation
|
|
self.get_config_name(k): v
|
|
for k, v in dict(mapping).items()
|
|
}
|
|
|
|
@classmethod
|
|
def from_str(cls, text):
|
|
"""
|
|
Build a :class:`TypedKernelConfig` out of the string content of a
|
|
Kconfig file.
|
|
"""
|
|
return cls(cls._parse_text(text))
|
|
|
|
@staticmethod
|
|
def _val_to_str(val):
|
|
"Convert back values to Kconfig-style string value"
|
|
# Special case the gracefully handle the output of get()
|
|
if val is None:
|
|
return None
|
|
elif isinstance(val, KernelConfigTristate):
|
|
return val.value
|
|
elif isinstance(val, basestring):
|
|
return '"{}"'.format(val.strip('"'))
|
|
else:
|
|
return str(val)
|
|
|
|
def __str__(self):
|
|
return '\n'.join(
|
|
'{}={}'.format(k, self._val_to_str(v))
|
|
for k, v in self.items()
|
|
)
|
|
|
|
@staticmethod
|
|
def _parse_val(k, v):
|
|
"""
|
|
Parse a value of types handled by Kconfig:
|
|
* string
|
|
* bool
|
|
* tristate
|
|
* hex
|
|
* int
|
|
|
|
Since bool cannot be distinguished from tristate, tristate is
|
|
always used. :meth:`KernelConfigTristate.__bool__` will allow using
|
|
it as a bool though, so it should not impact user code.
|
|
"""
|
|
if not v:
|
|
return None
|
|
|
|
# Handle "string" type
|
|
if v.startswith('"'):
|
|
# Strip enclosing "
|
|
return v[1:-1]
|
|
|
|
else:
|
|
try:
|
|
# Handles "bool" and "tristate" types
|
|
return KernelConfigTristate.from_str(v)
|
|
except ValueError:
|
|
pass
|
|
|
|
try:
|
|
# Handles "int" type
|
|
return int(v)
|
|
except ValueError:
|
|
pass
|
|
|
|
try:
|
|
# Handles "hex" type
|
|
return HexInt(v)
|
|
except ValueError:
|
|
pass
|
|
|
|
# If no type could be parsed
|
|
raise ValueError('Could not parse Kconfig key: {}={}'.format(
|
|
k, v
|
|
), k, v
|
|
)
|
|
|
|
@classmethod
|
|
def _parse_text(cls, text):
|
|
config = {}
|
|
for line in text.splitlines():
|
|
line = line.strip()
|
|
|
|
# skip empty lines
|
|
if not line:
|
|
continue
|
|
|
|
if line.startswith('#'):
|
|
match = cls.not_set_regex.search(line)
|
|
if match:
|
|
value = 'n'
|
|
name = match.group(1)
|
|
else:
|
|
continue
|
|
else:
|
|
name, value = line.split('=', 1)
|
|
|
|
name = cls.get_config_name(name.strip())
|
|
value = cls._parse_val(name, value.strip())
|
|
config[name] = value
|
|
return config
|
|
|
|
def __getitem__(self, name):
|
|
name = self.get_config_name(name)
|
|
try:
|
|
return self._config[name]
|
|
except KeyError:
|
|
raise KernelConfigKeyError(
|
|
"{} is not exposed in kernel config".format(name),
|
|
name
|
|
)
|
|
|
|
def __iter__(self):
|
|
return iter(self._config)
|
|
|
|
def __len__(self):
|
|
return len(self._config)
|
|
|
|
def __contains__(self, name):
|
|
name = self.get_config_name(name)
|
|
return name in self._config
|
|
|
|
def like(self, name):
|
|
regex = re.compile(name, re.I)
|
|
return {
|
|
k: v for k, v in self.items()
|
|
if regex.search(k)
|
|
}
|
|
|
|
def is_enabled(self, name):
|
|
return self.get(name) is KernelConfigTristate.YES
|
|
|
|
def is_module(self, name):
|
|
return self.get(name) is KernelConfigTristate.MODULE
|
|
|
|
def is_not_set(self, name):
|
|
return self.get(name) is KernelConfigTristate.NO
|
|
|
|
def has(self, name):
|
|
return self.is_enabled(name) or self.is_module(name)
|
|
|
|
|
|
class KernelConfig(object):
|
|
"""
|
|
Backward compatibility shim on top of :class:`TypedKernelConfig`.
|
|
|
|
This class does not provide a Mapping API and only return string values.
|
|
"""
|
|
@staticmethod
|
|
def get_config_name(name):
|
|
return TypedKernelConfig.get_config_name(name)
|
|
|
|
def __init__(self, text):
|
|
# Expose typed_config as a non-private attribute, so that user code
|
|
# needing it can get it from any existing producer of KernelConfig.
|
|
self.typed_config = TypedKernelConfig.from_str(text)
|
|
# Expose the original text for backward compatibility
|
|
self.text = text
|
|
|
|
def __bool__(self):
|
|
return bool(self.typed_config)
|
|
|
|
not_set_regex = TypedKernelConfig.not_set_regex
|
|
|
|
def iteritems(self):
|
|
for k, v in self.typed_config.items():
|
|
yield (k, self.typed_config._val_to_str(v))
|
|
|
|
items = iteritems
|
|
|
|
def get(self, name, strict=False):
|
|
if strict:
|
|
val = self.typed_config[name]
|
|
else:
|
|
val = self.typed_config.get(name)
|
|
|
|
return self.typed_config._val_to_str(val)
|
|
|
|
def like(self, name):
|
|
return {
|
|
k: self.typed_config._val_to_str(v)
|
|
for k, v in self.typed_config.like(name).items()
|
|
}
|
|
|
|
def is_enabled(self, name):
|
|
return self.typed_config.is_enabled(name)
|
|
|
|
def is_module(self, name):
|
|
return self.typed_config.is_module(name)
|
|
|
|
def is_not_set(self, name):
|
|
return self.typed_config.is_not_set(name)
|
|
|
|
def has(self, name):
|
|
return self.typed_config.has(name)
|
|
|
|
|
|
class LocalLinuxTarget(LinuxTarget):
|
|
|
|
def __init__(self,
|
|
connection_settings=None,
|
|
platform=None,
|
|
working_directory=None,
|
|
executables_directory=None,
|
|
connect=True,
|
|
modules=None,
|
|
load_default_modules=True,
|
|
shell_prompt=DEFAULT_SHELL_PROMPT,
|
|
conn_cls=LocalConnection,
|
|
is_container=False,
|
|
max_async=50,
|
|
):
|
|
super(LocalLinuxTarget, self).__init__(connection_settings=connection_settings,
|
|
platform=platform,
|
|
working_directory=working_directory,
|
|
executables_directory=executables_directory,
|
|
connect=connect,
|
|
modules=modules,
|
|
load_default_modules=load_default_modules,
|
|
shell_prompt=shell_prompt,
|
|
conn_cls=conn_cls,
|
|
is_container=is_container,
|
|
max_async=max_async)
|
|
|
|
def _resolve_paths(self):
|
|
if self.working_directory is None:
|
|
self.working_directory = '/tmp/devlib-target'
|
|
self._file_transfer_cache = self.path.join(self.working_directory, '.file-cache')
|
|
if self.executables_directory is None:
|
|
self.executables_directory = '/tmp/devlib-target/bin'
|
|
|
|
|
|
def _get_model_name(section):
|
|
name_string = section['model name']
|
|
parts = name_string.split('@')[0].strip().split()
|
|
return ' '.join([p for p in parts
|
|
if '(' not in p and p != 'CPU'])
|
|
|
|
|
|
def _get_part_name(section):
|
|
implementer = section.get('CPU implementer', '0x0')
|
|
part = section['CPU part']
|
|
variant = section.get('CPU variant', '0x0')
|
|
name = get_cpu_name(*list(map(integer, [implementer, part, variant])))
|
|
if name is None:
|
|
name = f'{implementer}/{part}/{variant}'
|
|
return name
|
|
|
|
|
|
def _build_path_tree(path_map, basepath, sep=os.path.sep, dictcls=dict):
|
|
"""
|
|
Convert a flat mapping of paths to values into a nested structure of
|
|
dict-line object (``dict``'s by default), mirroring the directory hierarchy
|
|
represented by the paths relative to ``basepath``.
|
|
|
|
"""
|
|
def process_node(node, path, value):
|
|
parts = path.split(sep, 1)
|
|
if len(parts) == 1: # leaf
|
|
node[parts[0]] = value
|
|
else: # branch
|
|
if parts[0] not in node:
|
|
node[parts[0]] = dictcls()
|
|
process_node(node[parts[0]], parts[1], value)
|
|
|
|
relpath_map = {os.path.relpath(p, basepath): v
|
|
for p, v in path_map.items()}
|
|
|
|
if len(relpath_map) == 1 and list(relpath_map.keys())[0] == '.':
|
|
result = list(relpath_map.values())[0]
|
|
else:
|
|
result = dictcls()
|
|
for path, value in relpath_map.items():
|
|
process_node(result, path, value)
|
|
|
|
return result
|
|
|
|
|
|
class ChromeOsTarget(LinuxTarget):
|
|
"""
|
|
Class for interacting with ChromeOS targets.
|
|
"""
|
|
|
|
os = 'chromeos'
|
|
|
|
# pylint: disable=too-many-locals,too-many-arguments
|
|
def __init__(self,
|
|
connection_settings=None,
|
|
platform=None,
|
|
working_directory=None,
|
|
executables_directory=None,
|
|
android_working_directory=None,
|
|
android_executables_directory=None,
|
|
connect=True,
|
|
modules=None,
|
|
load_default_modules=True,
|
|
shell_prompt=DEFAULT_SHELL_PROMPT,
|
|
package_data_directory="/data/data",
|
|
is_container=False,
|
|
max_async=50,
|
|
):
|
|
|
|
self.supports_android = None
|
|
self.android_container = None
|
|
|
|
# Pull out ssh connection settings
|
|
ssh_conn_params = ['host', 'username', 'password', 'keyfile',
|
|
'port', 'timeout', 'sudo_cmd',
|
|
'strict_host_check', 'use_scp',
|
|
'total_transfer_timeout', 'poll_transfers',
|
|
'start_transfer_poll_delay']
|
|
self.ssh_connection_settings = {}
|
|
self.ssh_connection_settings.update(
|
|
(key, value)
|
|
for key, value in connection_settings.items()
|
|
if key in ssh_conn_params
|
|
)
|
|
|
|
super().__init__(connection_settings=self.ssh_connection_settings,
|
|
platform=platform,
|
|
working_directory=working_directory,
|
|
executables_directory=executables_directory,
|
|
connect=False,
|
|
modules=modules,
|
|
load_default_modules=load_default_modules,
|
|
shell_prompt=shell_prompt,
|
|
conn_cls=SshConnection,
|
|
is_container=is_container,
|
|
max_async=max_async)
|
|
|
|
# We can't determine if the target supports android until connected to the linux host so
|
|
# create unconditionally.
|
|
# Pull out adb connection settings
|
|
adb_conn_params = ['device', 'adb_server', 'timeout']
|
|
self.android_connection_settings = {}
|
|
self.android_connection_settings.update(
|
|
(key, value)
|
|
for key, value in connection_settings.items()
|
|
if key in adb_conn_params
|
|
)
|
|
|
|
# If adb device is not explicitly specified use same as ssh host
|
|
if not connection_settings.get('device', None):
|
|
self.android_connection_settings['device'] = connection_settings.get('host', None)
|
|
|
|
self.android_container = AndroidTarget(connection_settings=self.android_connection_settings,
|
|
platform=platform,
|
|
working_directory=android_working_directory,
|
|
executables_directory=android_executables_directory,
|
|
connect=False,
|
|
load_default_modules=False,
|
|
shell_prompt=shell_prompt,
|
|
conn_cls=AdbConnection,
|
|
package_data_directory=package_data_directory,
|
|
is_container=True)
|
|
if connect:
|
|
self.connect()
|
|
|
|
def __getattr__(self, attr):
|
|
"""
|
|
By default use the linux target methods and attributes however,
|
|
if not present, use android implementation if available.
|
|
"""
|
|
try:
|
|
return super().__getattribute__(attr)
|
|
except AttributeError:
|
|
if hasattr(self.android_container, attr):
|
|
return getattr(self.android_container, attr)
|
|
raise
|
|
|
|
@asyn.asyncf
|
|
async def connect(self, timeout=30, check_boot_completed=True, max_async=None):
|
|
super().connect(
|
|
timeout=timeout,
|
|
check_boot_completed=check_boot_completed,
|
|
max_async=max_async,
|
|
)
|
|
|
|
# Assume device supports android apps if container directory is present
|
|
if self.supports_android is None:
|
|
self.supports_android = self.directory_exists('/opt/google/containers/android/')
|
|
|
|
if self.supports_android:
|
|
self.android_container.connect(timeout)
|
|
else:
|
|
self.android_container = None
|
|
|
|
def _resolve_paths(self):
|
|
if self.working_directory is None:
|
|
self.working_directory = '/mnt/stateful_partition/devlib-target'
|
|
self._file_transfer_cache = self.path.join(self.working_directory, '.file-cache')
|
|
if self.executables_directory is None:
|
|
self.executables_directory = self.path.join(self.working_directory, 'bin')
|