1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
|
# Copyright (C) 2017 Xilinx
# Based on externalsrc.bbclass, original copyrights follow:
# Copyright (C) 2012 Linux Foundation
# Some code and influence taken from srctree.bbclass:
# Copyright (C) 2009 Chris Larson <clarson@kergoth.com>
# Released under the MIT license (see COPYING.MIT for the terms)
SRCTREECOVEREDTASKS ?= "do_patch do_unpack do_fetch"
EXTERNALXSCTSRCHASH ?= "src build"
python () {
externalsrc = d.getVar('EXTERNALXSCTSRC')
if externalsrc:
import oe.recipeutils
import oe.path
d.setVar('BB_DONT_CACHE', '1')
d.setVar('S', externalsrc)
externalsrcbuild = d.getVar('EXTERNALXSCTSRC_BUILD')
if externalsrcbuild:
d.setVar('B', externalsrcbuild)
else:
d.setVar('B', '${WORKDIR}/${BPN}-${PV}/')
if d.getVar('S') == d.getVar('B'):
bb.error("Cannot set build directory to be same as source directory")
return None
local_srcuri = []
fetch = bb.fetch2.Fetch((d.getVar('SRC_URI') or '').split(), d)
for url in fetch.urls:
url_data = fetch.ud[url]
if (url_data.type == 'file'):
local_srcuri.append(url)
d.setVar('SRC_URI', ' '.join(local_srcuri))
if '{SRCPV}' in d.getVar('PV', False):
# Dummy value because the default function can't be called with blank SRC_URI
d.setVar('SRCPV', '999')
tasks = filter(lambda k: d.getVarFlag(k, "task", True), d.keys())
for task in tasks:
if task.endswith("_setscene"):
# sstate is never going to work for external source trees, disable it
bb.build.deltask(task, d)
else:
# Since configure will likely touch ${S}, ensure only we lock so one task has access at a time
d.appendVarFlag(task, "lockfiles", " ${TMPDIR}/singlexscttask.lock")
# We do not want our source to be wiped out, ever (kernel.bbclass does this for do_clean)
cleandirs = oe.recipeutils.split_var_value(d.getVarFlag(task, 'cleandirs', False) or '')
setvalue = False
for cleandir in cleandirs[:]:
if oe.path.is_path_parent(externalsrc, d.expand(cleandir)):
cleandirs.remove(cleandir)
setvalue = True
if setvalue:
d.setVarFlag(task, 'cleandirs', ' '.join(cleandirs))
fetch_tasks = ['do_fetch', 'do_unpack']
# If we deltask do_patch, there's no dependency to ensure do_unpack gets run, so add one
# Note that we cannot use d.appendVarFlag() here because deps is expected to be a list object, not a string
d.setVarFlag('do_configure', 'deps', (d.getVarFlag('do_configure', 'deps', False) or []) + ['do_unpack'])
for task in d.getVar("SRCTREECOVEREDTASKS").split():
if local_srcuri and task in fetch_tasks:
continue
bb.build.deltask(task, d)
d.prependVarFlag('do_compile', 'prefuncs', "xsct_externalsrc_compile_prefunc")
external_xsct_src_hash = d.getVar('EXTERNALXSCTSRCHASH')
# If EXTERNALXSCTSRCHASH is set to both build and src, hash both.
# If EXTERNALXSCTSRCHASH is set to build, hash build file. By default hash src files if EXTERNALXSCTSRCHASH is empty.
if ((external_xsct_src_hash.find('build') != -1) and (external_xsct_src_hash.find('src')!= -1)):
d.setVarFlag('do_compile', 'file-checksums', '${@xsct_buildtree_hash_files(d)} ${@xsct_srctree_hash_files(d)}')
elif external_xsct_src_hash.find('build') != -1:
d.setVarFlag('do_compile', 'file-checksums', '${@xsct_buildtree_hash_files(d)}')
else:
d.setVarFlag('do_compile', 'file-checksums', '${@xsct_srctree_hash_files(d)}')
# We don't want the workdir to go away
d.appendVar('RM_WORK_EXCLUDE', ' ' + d.getVar('PN'))
}
COMPILE_TRIGGER_FILES = "${XSCTH_WS}/${XSCTH_PROJ}"
python xsct_externalsrc_compile_prefunc() {
# Make it obvious that this is happening, since forgetting about it could lead to much confusion
bb.plain('NOTE: %s: compiling from external source tree %s' % (d.getVar('PN'), d.getVar('EXTERNALXSCTSRC')))
}
def xsct_srctree_hash_files(d):
import shutil
import subprocess
import tempfile
s_dir = d.getVar('EXTERNALXSCTSRC')
git_dir = os.path.join(s_dir, '.git')
oe_hash_file = os.path.join(git_dir, 'oe-devtool-tree-sha1')
ret = " "
if os.path.exists(git_dir):
with tempfile.NamedTemporaryFile(dir=git_dir, prefix='oe-devtool-index') as tmp_index:
# Clone index
shutil.copy2(os.path.join(git_dir, 'index'), tmp_index.name)
# Update our custom index
env = os.environ.copy()
env['GIT_INDEX_FILE'] = tmp_index.name
subprocess.check_output(['git', 'add', '.'], cwd=s_dir, env=env)
sha1 = subprocess.check_output(['git', 'write-tree'], cwd=s_dir, env=env).decode("utf-8")
with open(oe_hash_file, 'w') as fobj:
fobj.write(sha1)
ret = oe_hash_file + ':True'
else:
ret = d.getVar('EXTERNALXSCTSRC') + '/*:True'
return ret
def xsct_buildtree_hash_files(d):
import os
"""
Get the list of files that should trigger do_compile to re-execute,
"""
in_files = (d.getVar('COMPILE_TRIGGER_FILES') or '').split()
out_items = []
for entry in in_files:
if os.path.isdir(entry):
for path, subdirs, files in os.walk(entry):
for file in files:
if file.endswith(('.c', '.h', '.tcl')):
out_items.append('%s:True' % (os.path.join(path, file)))
else:
out_items.append('%s:%s' % (entry, os.path.exists(entry)))
return ' '.join(out_items)
|