2017-06-14 13:36:37 +02:00
|
|
|
# kas - setup tool for bitbake based projects
|
|
|
|
#
|
2018-09-05 11:13:03 +02:00
|
|
|
# Copyright (c) Siemens AG, 2017-2018
|
2017-06-14 13:36:37 +02:00
|
|
|
#
|
|
|
|
# Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
|
|
# of this software and associated documentation files (the "Software"), to deal
|
|
|
|
# in the Software without restriction, including without limitation the rights
|
|
|
|
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
|
|
# copies of the Software, and to permit persons to whom the Software is
|
|
|
|
# furnished to do so, subject to the following conditions:
|
|
|
|
#
|
|
|
|
# The above copyright notice and this permission notice shall be
|
|
|
|
# included in all copies or substantial portions of the Software.
|
|
|
|
#
|
|
|
|
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
|
|
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
|
|
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
|
|
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
|
|
# SOFTWARE.
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
|
|
|
This module contains the core implementation of kas.
|
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
import re
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
import logging
|
|
|
|
import tempfile
|
|
|
|
import asyncio
|
2018-01-25 19:07:39 +01:00
|
|
|
import errno
|
2017-06-14 13:36:37 +02:00
|
|
|
from subprocess import Popen, PIPE
|
2018-08-24 19:18:00 +02:00
|
|
|
from .context import get_context
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
__license__ = 'MIT'
|
2018-09-05 11:13:03 +02:00
|
|
|
__copyright__ = 'Copyright (c) Siemens AG, 2017-2018'
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
|
|
|
|
class LogOutput:
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
|
|
|
Handles the log output of executed applications
|
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
def __init__(self, live):
|
|
|
|
self.live = live
|
|
|
|
self.stdout = []
|
|
|
|
self.stderr = []
|
|
|
|
|
|
|
|
def log_stdout(self, line):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-09-05 11:13:03 +02:00
|
|
|
This method is called when a line is received over stdout.
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
if self.live:
|
|
|
|
logging.info(line.strip())
|
|
|
|
self.stdout.append(line)
|
|
|
|
|
|
|
|
def log_stderr(self, line):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-09-05 11:13:03 +02:00
|
|
|
This method is called when a line is received over stderr.
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
if self.live:
|
|
|
|
logging.error(line.strip())
|
|
|
|
self.stderr.append(line)
|
|
|
|
|
|
|
|
|
|
|
|
@asyncio.coroutine
|
2017-06-21 13:32:56 +02:00
|
|
|
def _read_stream(stream, callback):
|
|
|
|
"""
|
2017-06-26 16:07:53 +02:00
|
|
|
This asynchronous method reads from the output stream of the
|
2017-06-21 13:32:56 +02:00
|
|
|
application and transfers each line to the callback function.
|
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
while True:
|
|
|
|
line = yield from stream.readline()
|
|
|
|
try:
|
|
|
|
line = line.decode('utf-8')
|
2017-06-21 13:32:56 +02:00
|
|
|
except UnicodeDecodeError as err:
|
2018-09-05 11:13:03 +02:00
|
|
|
logging.warning('Could not decode line from stream, ignoring: %s',
|
2017-06-21 13:32:56 +02:00
|
|
|
err)
|
2017-06-14 13:36:37 +02:00
|
|
|
if line:
|
2017-06-21 13:32:56 +02:00
|
|
|
callback(line)
|
2017-06-14 13:36:37 +02:00
|
|
|
else:
|
|
|
|
break
|
|
|
|
|
2017-06-21 13:32:55 +02:00
|
|
|
|
2017-06-14 13:36:37 +02:00
|
|
|
@asyncio.coroutine
|
2018-09-03 10:01:58 +02:00
|
|
|
def run_cmd_async(cmd, cwd, env=None, fail=True, liveupdate=True):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2017-06-28 12:43:55 +02:00
|
|
|
Run a command asynchronously.
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-07-18 15:15:12 +02:00
|
|
|
|
2018-08-24 19:18:00 +02:00
|
|
|
env = env or get_context().environ
|
2018-09-03 10:01:58 +02:00
|
|
|
cmdstr = ' '.join(cmd)
|
2017-06-28 12:43:55 +02:00
|
|
|
logging.info('%s$ %s', cwd, cmdstr)
|
|
|
|
|
|
|
|
logo = LogOutput(liveupdate)
|
|
|
|
|
2018-01-25 19:07:39 +01:00
|
|
|
try:
|
2018-09-03 10:01:58 +02:00
|
|
|
process = yield from asyncio.create_subprocess_exec(
|
|
|
|
*cmd,
|
|
|
|
cwd=cwd,
|
|
|
|
env=env,
|
|
|
|
stdout=asyncio.subprocess.PIPE,
|
|
|
|
stderr=asyncio.subprocess.PIPE)
|
2018-01-25 19:07:39 +01:00
|
|
|
except FileNotFoundError as ex:
|
|
|
|
if fail:
|
|
|
|
raise ex
|
|
|
|
return (errno.ENOENT, str(ex))
|
|
|
|
except PermissionError as ex:
|
|
|
|
if fail:
|
|
|
|
raise ex
|
|
|
|
return (errno.EPERM, str(ex))
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
yield from asyncio.wait([
|
2017-06-28 12:43:55 +02:00
|
|
|
_read_stream(process.stdout, logo.log_stdout),
|
|
|
|
_read_stream(process.stderr, logo.log_stderr)
|
2017-06-14 13:36:37 +02:00
|
|
|
])
|
|
|
|
ret = yield from process.wait()
|
2017-06-28 12:43:55 +02:00
|
|
|
|
|
|
|
if ret and fail:
|
2017-06-28 19:25:59 +02:00
|
|
|
msg = 'Command "{cwd}$ {cmd}" failed'.format(cwd=cwd, cmd=cmdstr)
|
2017-07-07 12:15:54 +02:00
|
|
|
if logo.stderr:
|
|
|
|
msg += '\n--- Error summary ---\n'
|
|
|
|
for line in logo.stderr:
|
|
|
|
msg += line
|
2017-06-28 12:43:55 +02:00
|
|
|
logging.error(msg)
|
|
|
|
|
|
|
|
return (ret, ''.join(logo.stdout))
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
|
2018-09-03 10:01:58 +02:00
|
|
|
def run_cmd(cmd, cwd, env=None, fail=True, liveupdate=True):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2017-06-28 12:43:55 +02:00
|
|
|
Runs a command synchronously.
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
|
|
|
|
2017-06-28 12:43:52 +02:00
|
|
|
loop = asyncio.get_event_loop()
|
2017-06-28 19:26:00 +02:00
|
|
|
(ret, output) = loop.run_until_complete(
|
2018-09-03 10:01:58 +02:00
|
|
|
run_cmd_async(cmd, cwd, env, fail, liveupdate))
|
2017-06-28 19:26:00 +02:00
|
|
|
if ret and fail:
|
|
|
|
sys.exit(ret)
|
|
|
|
return (ret, output)
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
|
|
|
|
def find_program(paths, name):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
|
|
|
Find a file within the paths array and returns its path.
|
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
for path in paths.split(os.pathsep):
|
|
|
|
prg = os.path.join(path, name)
|
|
|
|
if os.path.isfile(prg):
|
|
|
|
return prg
|
|
|
|
return None
|
|
|
|
|
|
|
|
|
2018-11-07 14:48:31 +01:00
|
|
|
def _create_task(routine):
|
|
|
|
try:
|
|
|
|
creation_func = asyncio.ensure_future
|
|
|
|
except AttributeError:
|
|
|
|
# for Python < 3.5, avoiding the keyword 'async' introduced in 3.7
|
|
|
|
creation_func = getattr(asyncio, 'async')
|
|
|
|
|
|
|
|
return creation_func(routine)
|
|
|
|
|
|
|
|
|
2018-08-24 19:18:06 +02:00
|
|
|
def repos_fetch(repos):
|
2017-06-28 12:43:57 +02:00
|
|
|
"""
|
|
|
|
Fetches the list of repositories to the kas_work_dir.
|
|
|
|
"""
|
2017-07-03 11:34:32 +02:00
|
|
|
tasks = []
|
2017-06-28 12:43:57 +02:00
|
|
|
for repo in repos:
|
2018-11-07 14:48:31 +01:00
|
|
|
tasks.append(_create_task(repo.fetch_async()))
|
2017-06-28 12:43:57 +02:00
|
|
|
|
|
|
|
loop = asyncio.get_event_loop()
|
2017-07-03 11:34:32 +02:00
|
|
|
loop.run_until_complete(asyncio.wait(tasks))
|
|
|
|
|
|
|
|
for task in tasks:
|
|
|
|
if task.result():
|
|
|
|
sys.exit(task.result())
|
2017-06-22 12:28:32 +02:00
|
|
|
|
|
|
|
|
2018-08-24 19:18:06 +02:00
|
|
|
def repos_apply_patches(repos):
|
2018-03-09 08:22:52 +01:00
|
|
|
"""
|
|
|
|
Applies the patches to the repositories.
|
|
|
|
"""
|
|
|
|
tasks = []
|
|
|
|
for repo in repos:
|
2018-11-07 14:48:31 +01:00
|
|
|
tasks.append(_create_task(repo.apply_patches_async()))
|
2018-03-09 08:22:52 +01:00
|
|
|
|
|
|
|
loop = asyncio.get_event_loop()
|
|
|
|
loop.run_until_complete(asyncio.wait(tasks))
|
|
|
|
|
|
|
|
for task in tasks:
|
|
|
|
if task.result():
|
|
|
|
sys.exit(task.result())
|
|
|
|
|
|
|
|
|
2018-08-24 19:18:05 +02:00
|
|
|
def get_build_environ():
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-09-05 11:13:03 +02:00
|
|
|
Creates the build environment variables.
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2017-06-22 10:29:10 +02:00
|
|
|
# nasty side effect function: running oe/isar-init-build-env also
|
2017-06-14 13:36:37 +02:00
|
|
|
# creates the conf directory
|
|
|
|
|
2017-07-28 17:28:56 +02:00
|
|
|
init_repo = None
|
2017-06-22 10:29:10 +02:00
|
|
|
permutations = \
|
2018-08-24 19:18:05 +02:00
|
|
|
[(repo, script) for repo in get_context().config.get_repos()
|
2017-06-22 10:29:10 +02:00
|
|
|
for script in ['oe-init-build-env', 'isar-init-build-env']]
|
|
|
|
for (repo, script) in permutations:
|
|
|
|
if os.path.exists(repo.path + '/' + script):
|
2017-07-28 17:28:56 +02:00
|
|
|
if init_repo:
|
|
|
|
logging.error('Multiple init scripts found (%s vs. %s). ',
|
|
|
|
repo.name, init_repo.name)
|
|
|
|
logging.error('Resolve ambiguity by removing one of the repos')
|
|
|
|
sys.exit(1)
|
|
|
|
init_repo = repo
|
2017-06-22 10:29:10 +02:00
|
|
|
init_script = script
|
2017-07-28 17:28:56 +02:00
|
|
|
if not init_repo:
|
2017-06-22 10:29:10 +02:00
|
|
|
logging.error('Did not find any init-build-env script')
|
2017-06-14 13:36:37 +02:00
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
get_bb_env_file = tempfile.mktemp()
|
2017-06-21 13:32:56 +02:00
|
|
|
with open(get_bb_env_file, 'w') as fds:
|
2017-06-14 13:36:37 +02:00
|
|
|
script = """#!/bin/bash
|
2017-07-28 16:29:44 +02:00
|
|
|
set -e
|
|
|
|
source %s $1 > /dev/null
|
2017-06-14 13:36:37 +02:00
|
|
|
env
|
2017-06-22 10:29:10 +02:00
|
|
|
""" % init_script
|
2017-06-21 13:32:56 +02:00
|
|
|
fds.write(script)
|
2017-06-14 13:36:37 +02:00
|
|
|
os.chmod(get_bb_env_file, 0o775)
|
|
|
|
|
|
|
|
env = {}
|
2017-11-16 15:45:39 +01:00
|
|
|
env['PATH'] = '/usr/sbin:/usr/bin:/sbin:/bin'
|
2017-06-14 13:36:37 +02:00
|
|
|
|
2018-08-24 19:18:05 +02:00
|
|
|
(_, output) = run_cmd([get_bb_env_file, get_context().build_dir],
|
2017-07-28 17:28:56 +02:00
|
|
|
cwd=init_repo.path, env=env, liveupdate=False)
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
os.remove(get_bb_env_file)
|
|
|
|
|
|
|
|
env = {}
|
|
|
|
for line in output.splitlines():
|
|
|
|
try:
|
|
|
|
(key, val) = line.split('=', 1)
|
|
|
|
env[key] = val
|
2017-06-21 13:32:56 +02:00
|
|
|
except ValueError:
|
2017-06-14 13:36:37 +02:00
|
|
|
pass
|
|
|
|
|
2018-08-24 19:18:05 +02:00
|
|
|
conf_env = get_context().config.get_environment()
|
2017-11-23 14:15:51 +01:00
|
|
|
|
2017-06-21 13:32:56 +02:00
|
|
|
env_vars = ['SSTATE_DIR', 'DL_DIR', 'TMPDIR']
|
2017-11-23 14:15:51 +01:00
|
|
|
env_vars.extend(conf_env)
|
|
|
|
|
|
|
|
env.update(conf_env)
|
|
|
|
|
2017-06-14 13:36:37 +02:00
|
|
|
if 'BB_ENV_EXTRAWHITE' in env:
|
2017-12-22 09:51:16 +01:00
|
|
|
extra_white = env['BB_ENV_EXTRAWHITE'] + ' ' + ' '.join(env_vars)
|
2017-06-21 13:32:56 +02:00
|
|
|
env.update({'BB_ENV_EXTRAWHITE': extra_white})
|
2017-06-14 13:36:37 +02:00
|
|
|
|
2018-09-14 19:27:47 +02:00
|
|
|
env_vars.extend(['SSH_AUTH_SOCK',
|
2017-07-19 08:51:19 +02:00
|
|
|
'SHELL', 'TERM',
|
|
|
|
'GIT_PROXY_COMMAND', 'NO_PROXY'])
|
2017-06-19 10:11:35 +02:00
|
|
|
|
2017-06-21 13:32:56 +02:00
|
|
|
for env_var in env_vars:
|
|
|
|
if env_var in os.environ:
|
|
|
|
env[env_var] = os.environ[env_var]
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
return env
|
|
|
|
|
|
|
|
|
|
|
|
def ssh_add_key(env, key):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-09-05 11:13:03 +02:00
|
|
|
Adds an ssh key to the ssh-agent
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2017-09-08 17:09:22 +02:00
|
|
|
process = Popen(['ssh-add', '-'], stdin=PIPE, stdout=None,
|
2017-06-21 13:32:56 +02:00
|
|
|
stderr=PIPE, env=env)
|
|
|
|
(_, error) = process.communicate(input=str.encode(key))
|
|
|
|
if process.returncode and error:
|
|
|
|
logging.error('failed to add ssh key: %s', error)
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
|
2018-08-24 19:18:04 +02:00
|
|
|
def ssh_cleanup_agent():
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-09-05 11:13:03 +02:00
|
|
|
Removes the identities and stops the ssh-agent instance
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
2018-08-24 19:18:04 +02:00
|
|
|
env = get_context().environ
|
2017-06-14 13:36:37 +02:00
|
|
|
# remove the identities
|
2018-08-24 19:18:04 +02:00
|
|
|
process = Popen(['ssh-add', '-D'], env=env)
|
2017-06-21 13:32:56 +02:00
|
|
|
process.wait()
|
|
|
|
if process.returncode != 0:
|
2017-06-14 13:36:37 +02:00
|
|
|
logging.error('failed to delete SSH identities')
|
|
|
|
|
|
|
|
# stop the ssh-agent
|
2018-08-24 19:18:04 +02:00
|
|
|
process = Popen(['ssh-agent', '-k'], env=env)
|
2017-06-21 13:32:56 +02:00
|
|
|
process.wait()
|
|
|
|
if process.returncode != 0:
|
2017-06-14 13:36:37 +02:00
|
|
|
logging.error('failed to stop SSH agent')
|
|
|
|
|
|
|
|
|
2018-08-24 19:18:04 +02:00
|
|
|
def ssh_setup_agent(envkeys=None):
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
|
|
|
Starts the ssh-agent
|
|
|
|
"""
|
2018-08-24 19:18:04 +02:00
|
|
|
env = get_context().environ
|
2017-06-21 13:32:56 +02:00
|
|
|
envkeys = envkeys or ['SSH_PRIVATE_KEY']
|
2017-09-08 17:09:22 +02:00
|
|
|
output = os.popen('ssh-agent -s').readlines()
|
2017-06-14 13:36:37 +02:00
|
|
|
for line in output:
|
2017-06-21 13:32:56 +02:00
|
|
|
matches = re.search(r"(\S+)\=(\S+)\;", line)
|
2017-06-14 13:36:37 +02:00
|
|
|
if matches:
|
2018-08-24 19:18:04 +02:00
|
|
|
env[matches.group(1)] = matches.group(2)
|
2017-06-14 13:36:37 +02:00
|
|
|
|
2017-06-21 13:32:56 +02:00
|
|
|
for envkey in envkeys:
|
|
|
|
key = os.environ.get(envkey)
|
2017-06-14 13:36:37 +02:00
|
|
|
if key:
|
2018-08-13 14:11:22 +02:00
|
|
|
logging.info("adding SSH key")
|
2018-08-24 19:18:04 +02:00
|
|
|
ssh_add_key(env, key)
|
2017-06-14 13:36:37 +02:00
|
|
|
else:
|
2017-06-21 13:32:56 +02:00
|
|
|
logging.warning('%s is missing', envkey)
|
2017-06-14 13:36:37 +02:00
|
|
|
|
|
|
|
|
2018-08-24 19:18:04 +02:00
|
|
|
def ssh_no_host_key_check():
|
2017-06-21 13:32:56 +02:00
|
|
|
"""
|
|
|
|
Disables ssh host key check
|
|
|
|
"""
|
2017-06-14 13:36:37 +02:00
|
|
|
home = os.path.expanduser('~')
|
|
|
|
if not os.path.exists(home + '/.ssh'):
|
|
|
|
os.mkdir(home + '/.ssh')
|
2017-06-21 13:32:56 +02:00
|
|
|
with open(home + '/.ssh/config', 'w') as fds:
|
|
|
|
fds.write('Host *\n\tStrictHostKeyChecking no\n\n')
|
2017-06-28 14:48:41 +02:00
|
|
|
|
|
|
|
|
|
|
|
def kasplugin(plugin_class):
|
|
|
|
"""
|
2018-09-05 11:13:03 +02:00
|
|
|
A decorator that registers kas plugins
|
2017-06-28 14:48:41 +02:00
|
|
|
"""
|
|
|
|
if not hasattr(kasplugin, 'plugins'):
|
|
|
|
setattr(kasplugin, 'plugins', [])
|
|
|
|
getattr(kasplugin, 'plugins').append(plugin_class)
|