classes: Remove and sanitise import statements

Signed-off-by: Richard Purdie <rpurdie@linux.intel.com>
This commit is contained in:
Richard Purdie 2009-11-08 22:32:43 +00:00
parent 80d55bbd6e
commit c54117458a
31 changed files with 34 additions and 147 deletions

View File

@ -2,8 +2,6 @@
AUTOTOOLS_NATIVE_STAGE_INSTALL = "1" AUTOTOOLS_NATIVE_STAGE_INSTALL = "1"
def autotools_dep_prepend(d): def autotools_dep_prepend(d):
import bb;
if bb.data.getVar('INHIBIT_AUTOTOOLS_DEPS', d, 1): if bb.data.getVar('INHIBIT_AUTOTOOLS_DEPS', d, 1):
return '' return ''
@ -29,7 +27,6 @@ acpaths = "default"
EXTRA_AUTORECONF = "--exclude=autopoint" EXTRA_AUTORECONF = "--exclude=autopoint"
def autotools_set_crosscompiling(d): def autotools_set_crosscompiling(d):
import bb
if not bb.data.inherits_class('native', d): if not bb.data.inherits_class('native', d):
return " cross_compiling=yes" return " cross_compiling=yes"
return "" return ""

View File

@ -1,6 +1,5 @@
# The namespaces can clash here hence the two step replace # The namespaces can clash here hence the two step replace
def get_binconfig_mangle(d): def get_binconfig_mangle(d):
import bb.data
s = "-e ''" s = "-e ''"
if not bb.data.inherits_class('native', d): if not bb.data.inherits_class('native', d):
optional_quote = r"\(\"\?\)" optional_quote = r"\(\"\?\)"

View File

@ -9,7 +9,7 @@ RDEPENDS += "${@["perl", ""][(bb.data.inherits_class('native', d))]}"
# Determine the staged version of perl from the perl configuration file # Determine the staged version of perl from the perl configuration file
def get_perl_version(d): def get_perl_version(d):
import os, bb, re import re
cfg = bb.data.expand('${STAGING_DIR}/${HOST_SYS}/perl/config.sh', d) cfg = bb.data.expand('${STAGING_DIR}/${HOST_SYS}/perl/config.sh', d)
try: try:
f = open(cfg, 'r') f = open(cfg, 'r')
@ -33,7 +33,6 @@ def is_new_perl(d):
# Determine where the library directories are # Determine where the library directories are
def perl_get_libdirs(d): def perl_get_libdirs(d):
import bb
libdir = bb.data.getVar('libdir', d, 1) libdir = bb.data.getVar('libdir', d, 1)
if is_new_perl(d) == "yes": if is_new_perl(d) == "yes":
libdirs = libdir + '/perl5' libdirs = libdir + '/perl5'
@ -42,7 +41,6 @@ def perl_get_libdirs(d):
return libdirs return libdirs
def is_target(d): def is_target(d):
import bb
if not bb.data.inherits_class('native', d): if not bb.data.inherits_class('native', d):
return "yes" return "yes"
return "no" return "no"

View File

@ -10,7 +10,6 @@ inherit cpan-base
# libmodule-build-perl) # libmodule-build-perl)
# #
def cpan_build_dep_prepend(d): def cpan_build_dep_prepend(d):
import bb;
if bb.data.getVar('CPAN_BUILD_DEPS', d, 1): if bb.data.getVar('CPAN_BUILD_DEPS', d, 1):
return '' return ''
pn = bb.data.getVar('PN', d, 1) pn = bb.data.getVar('PN', d, 1)

View File

@ -3,7 +3,6 @@ DEPENDS += "${@["python-native python", ""][(bb.data.getVar('PACKAGES', d, 1) =
RDEPENDS += "python-core" RDEPENDS += "python-core"
def python_dir(d): def python_dir(d):
import os, bb
staging_incdir = bb.data.getVar( "STAGING_INCDIR", d, 1 ) staging_incdir = bb.data.getVar( "STAGING_INCDIR", d, 1 )
if os.path.exists( "%s/python2.5" % staging_incdir ): return "python2.5" if os.path.exists( "%s/python2.5" % staging_incdir ): return "python2.5"
if os.path.exists( "%s/python2.4" % staging_incdir ): return "python2.4" if os.path.exists( "%s/python2.4" % staging_incdir ): return "python2.4"

View File

@ -1,5 +1,4 @@
def gettext_after_parse(d): def gettext_after_parse(d):
import bb
# Remove the NLS bits if USE_NLS is no. # Remove the NLS bits if USE_NLS is no.
if bb.data.getVar('USE_NLS', d, 1) == 'no': if bb.data.getVar('USE_NLS', d, 1) == 'no':
cfg = oe_filter_out('^--(dis|en)able-nls$', bb.data.getVar('EXTRA_OECONF', d, 1) or "", d) cfg = oe_filter_out('^--(dis|en)able-nls$', bb.data.getVar('EXTRA_OECONF', d, 1) or "", d)

View File

@ -33,7 +33,6 @@ def icc_determine_gcc_version(gcc):
'i686-apple-darwin8-gcc-4.0.1 (GCC) 4.0.1 (Apple Computer, Inc. build 5363)' 'i686-apple-darwin8-gcc-4.0.1 (GCC) 4.0.1 (Apple Computer, Inc. build 5363)'
""" """
import os
return os.popen("%s --version" % gcc ).readline().split()[2] return os.popen("%s --version" % gcc ).readline().split()[2]
def create_cross_env(bb,d): def create_cross_env(bb,d):
@ -47,7 +46,7 @@ def create_cross_env(bb,d):
if len(prefix) == 0: if len(prefix) == 0:
return "" return ""
import tarfile, socket, time, os import tarfile, socket, time
ice_dir = bb.data.expand('${CROSS_DIR}', d) ice_dir = bb.data.expand('${CROSS_DIR}', d)
prefix = bb.data.expand('${HOST_PREFIX}' , d) prefix = bb.data.expand('${HOST_PREFIX}' , d)
distro = bb.data.expand('${DISTRO}', d) distro = bb.data.expand('${DISTRO}', d)
@ -96,7 +95,7 @@ def create_cross_env(bb,d):
def create_native_env(bb,d): def create_native_env(bb,d):
import tarfile, socket, time, os import tarfile, socket, time
ice_dir = bb.data.expand('${CROSS_DIR}', d) ice_dir = bb.data.expand('${CROSS_DIR}', d)
prefix = bb.data.expand('${HOST_PREFIX}' , d) prefix = bb.data.expand('${HOST_PREFIX}' , d)
distro = bb.data.expand('${DISTRO}', d) distro = bb.data.expand('${DISTRO}', d)
@ -137,7 +136,7 @@ def create_native_env(bb,d):
def create_cross_kernel_env(bb,d): def create_cross_kernel_env(bb,d):
import tarfile, socket, time, os import tarfile, socket, time
ice_dir = bb.data.expand('${CROSS_DIR}', d) ice_dir = bb.data.expand('${CROSS_DIR}', d)
prefix = bb.data.expand('${HOST_PREFIX}' , d) prefix = bb.data.expand('${HOST_PREFIX}' , d)
distro = bb.data.expand('${DISTRO}', d) distro = bb.data.expand('${DISTRO}', d)
@ -204,8 +203,6 @@ def create_path(compilers, type, bb, d):
""" """
Create Symlinks for the icecc in the staging directory Create Symlinks for the icecc in the staging directory
""" """
import os
staging = os.path.join(bb.data.expand('${STAGING_DIR}', d), "ice", type) staging = os.path.join(bb.data.expand('${STAGING_DIR}', d), "ice", type)
#check if the icecc path is set by the user #check if the icecc path is set by the user

View File

@ -26,8 +26,6 @@ PACKAGE_ARCH = "${MACHINE_ARCH}"
do_rootfs[depends] += "makedevs-native:do_populate_staging fakeroot-native:do_populate_staging ldconfig-native:do_populate_staging" do_rootfs[depends] += "makedevs-native:do_populate_staging fakeroot-native:do_populate_staging ldconfig-native:do_populate_staging"
python () { python () {
import bb
deps = bb.data.getVarFlag('do_rootfs', 'depends', d) or "" deps = bb.data.getVarFlag('do_rootfs', 'depends', d) or ""
for type in (bb.data.getVar('IMAGE_FSTYPES', d, True) or "").split(): for type in (bb.data.getVar('IMAGE_FSTYPES', d, True) or "").split():
for dep in ((bb.data.getVar('IMAGE_DEPENDS_%s' % type, d) or "").split() or []): for dep in ((bb.data.getVar('IMAGE_DEPENDS_%s' % type, d) or "").split() or []):
@ -49,7 +47,6 @@ python () {
# is searched for in the BBPATH (same as the old version.) # is searched for in the BBPATH (same as the old version.)
# #
def get_devtable_list(d): def get_devtable_list(d):
import bb
devtable = bb.data.getVar('IMAGE_DEVICE_TABLE', d, 1) devtable = bb.data.getVar('IMAGE_DEVICE_TABLE', d, 1)
if devtable != None: if devtable != None:
return devtable return devtable
@ -62,7 +59,6 @@ def get_devtable_list(d):
return str return str
def get_imagecmds(d): def get_imagecmds(d):
import bb
cmds = "\n" cmds = "\n"
old_overrides = bb.data.getVar('OVERRIDES', d, 0) old_overrides = bb.data.getVar('OVERRIDES', d, 0)
for type in bb.data.getVar('IMAGE_FSTYPES', d, True).split(): for type in bb.data.getVar('IMAGE_FSTYPES', d, True).split():

View File

@ -169,7 +169,6 @@ def package_qa_get_elf(path, bits32):
def package_qa_clean_path(path,d): def package_qa_clean_path(path,d):
""" Remove the common prefix from the path. In this case it is the TMPDIR""" """ Remove the common prefix from the path. In this case it is the TMPDIR"""
import bb
return path.replace(bb.data.getVar('TMPDIR',d,True),"") return path.replace(bb.data.getVar('TMPDIR',d,True),"")
def package_qa_make_fatal_error(error_class, name, path,d): def package_qa_make_fatal_error(error_class, name, path,d):
@ -184,7 +183,6 @@ def package_qa_write_error(error_class, name, path, d):
""" """
Log the error Log the error
""" """
import bb, os
ERROR_NAMES =[ ERROR_NAMES =[
"non dev contains .so", "non dev contains .so",
@ -214,7 +212,6 @@ def package_qa_write_error(error_class, name, path, d):
f.close() f.close()
def package_qa_handle_error(error_class, error_msg, name, path, d): def package_qa_handle_error(error_class, error_msg, name, path, d):
import bb
fatal = package_qa_make_fatal_error(error_class, name, path, d) fatal = package_qa_make_fatal_error(error_class, name, path, d)
if fatal: if fatal:
bb.error("QA Issue: %s" % error_msg) bb.error("QA Issue: %s" % error_msg)
@ -229,7 +226,6 @@ def package_qa_check_rpath(file,name,d):
""" """
Check for dangerous RPATHs Check for dangerous RPATHs
""" """
import bb, os
sane = True sane = True
scanelf = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'scanelf') scanelf = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'scanelf')
bad_dir = bb.data.getVar('TMPDIR', d, True) + "/work" bad_dir = bb.data.getVar('TMPDIR', d, True) + "/work"
@ -255,7 +251,6 @@ def package_qa_check_devdbg(path, name,d):
non dev packages containing non dev packages containing
""" """
import bb, os
sane = True sane = True
if not "-dev" in name: if not "-dev" in name:
@ -283,7 +278,6 @@ def package_qa_check_arch(path,name,d):
""" """
Check if archs are compatible Check if archs are compatible
""" """
import bb, os
sane = True sane = True
target_os = bb.data.getVar('TARGET_OS', d, True) target_os = bb.data.getVar('TARGET_OS', d, True)
target_arch = bb.data.getVar('TARGET_ARCH', d, True) target_arch = bb.data.getVar('TARGET_ARCH', d, True)
@ -322,7 +316,6 @@ def package_qa_check_desktop(path, name, d):
""" """
Run all desktop files through desktop-file-validate. Run all desktop files through desktop-file-validate.
""" """
import bb, os
sane = True sane = True
if path.endswith(".desktop"): if path.endswith(".desktop"):
desktop_file_validate = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'desktop-file-validate') desktop_file_validate = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'desktop-file-validate')
@ -337,7 +330,6 @@ def package_qa_check_buildpaths(path, name, d):
""" """
Check for build paths inside target files and error if not found in the whitelist Check for build paths inside target files and error if not found in the whitelist
""" """
import bb, os
sane = True sane = True
# Ignore .debug files, not interesting # Ignore .debug files, not interesting
@ -364,7 +356,6 @@ def package_qa_check_staged(path,d):
to find the one responsible for the errors easily even to find the one responsible for the errors easily even
if we look at every .pc and .la file if we look at every .pc and .la file
""" """
import os, bb
sane = True sane = True
tmpdir = bb.data.getVar('TMPDIR', d, True) tmpdir = bb.data.getVar('TMPDIR', d, True)
@ -402,7 +393,6 @@ def package_qa_check_staged(path,d):
# Walk over all files in a directory and call func # Walk over all files in a directory and call func
def package_qa_walk(path, funcs, package,d): def package_qa_walk(path, funcs, package,d):
import os
sane = True sane = True
for root, dirs, files in os.walk(path): for root, dirs, files in os.walk(path):
@ -415,7 +405,6 @@ def package_qa_walk(path, funcs, package,d):
return sane return sane
def package_qa_check_rdepends(pkg, workdir, d): def package_qa_check_rdepends(pkg, workdir, d):
import bb
sane = True sane = True
if not "-dbg" in pkg and not "task-" in pkg and not "-image" in pkg: if not "-dbg" in pkg and not "task-" in pkg and not "-image" in pkg:
# Copied from package_ipk.bbclass # Copied from package_ipk.bbclass
@ -496,7 +485,6 @@ python do_qa_staging() {
addtask qa_configure after do_configure before do_compile addtask qa_configure after do_configure before do_compile
python do_qa_configure() { python do_qa_configure() {
bb.note("Checking sanity of the config.log file") bb.note("Checking sanity of the config.log file")
import os
for root, dirs, files in os.walk(bb.data.getVar('WORKDIR', d, True)): for root, dirs, files in os.walk(bb.data.getVar('WORKDIR', d, True)):
statement = "grep 'CROSS COMPILE Badness:' %s > /dev/null" % \ statement = "grep 'CROSS COMPILE Badness:' %s > /dev/null" % \
os.path.join(root,"config.log") os.path.join(root,"config.log")

View File

@ -15,7 +15,7 @@ valid_archs = "alpha cris ia64 \
avr32 blackfin" avr32 blackfin"
def map_kernel_arch(a, d): def map_kernel_arch(a, d):
import bb, re import re
valid_archs = bb.data.getVar('valid_archs', d, 1).split() valid_archs = bb.data.getVar('valid_archs', d, 1).split()

View File

@ -9,9 +9,6 @@ INHIBIT_DEFAULT_DEPS = "1"
KERNEL_IMAGETYPE ?= "zImage" KERNEL_IMAGETYPE ?= "zImage"
python __anonymous () { python __anonymous () {
import bb
kerneltype = bb.data.getVar('KERNEL_IMAGETYPE', d, 1) or '' kerneltype = bb.data.getVar('KERNEL_IMAGETYPE', d, 1) or ''
if kerneltype == 'uImage': if kerneltype == 'uImage':
depends = bb.data.getVar("DEPENDS", d, 1) depends = bb.data.getVar("DEPENDS", d, 1)
@ -271,7 +268,7 @@ module_conf_rfcomm = "alias bt-proto-3 rfcomm"
python populate_packages_prepend () { python populate_packages_prepend () {
def extract_modinfo(file): def extract_modinfo(file):
import os, re import re
tmpfile = os.tmpnam() tmpfile = os.tmpnam()
cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (bb.data.getVar("PATH", d, 1), bb.data.getVar("HOST_PREFIX", d, 1) or "", file, tmpfile) cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (bb.data.getVar("PATH", d, 1), bb.data.getVar("HOST_PREFIX", d, 1) or "", file, tmpfile)
os.system(cmd) os.system(cmd)
@ -289,7 +286,7 @@ python populate_packages_prepend () {
return vals return vals
def parse_depmod(): def parse_depmod():
import os, re import re
dvar = bb.data.getVar('D', d, 1) dvar = bb.data.getVar('D', d, 1)
if not dvar: if not dvar:
@ -343,7 +340,7 @@ python populate_packages_prepend () {
file = file.replace(bb.data.getVar('D', d, 1) or '', '', 1) file = file.replace(bb.data.getVar('D', d, 1) or '', '', 1)
if module_deps.has_key(file): if module_deps.has_key(file):
import os.path, re import re
dependencies = [] dependencies = []
for i in module_deps[file]: for i in module_deps[file]:
m = re.match(pattern, os.path.basename(i)) m = re.match(pattern, os.path.basename(i))
@ -411,7 +408,7 @@ python populate_packages_prepend () {
postrm = bb.data.getVar('pkg_postrm_modules', d, 1) postrm = bb.data.getVar('pkg_postrm_modules', d, 1)
do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % bb.data.getVar("KERNEL_VERSION", d, 1)) do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % bb.data.getVar("KERNEL_VERSION", d, 1))
import re, os import re
metapkg = "kernel-modules" metapkg = "kernel-modules"
bb.data.setVar('ALLOW_EMPTY_' + metapkg, "1", d) bb.data.setVar('ALLOW_EMPTY_' + metapkg, "1", d)
bb.data.setVar('FILES_' + metapkg, "", d) bb.data.setVar('FILES_' + metapkg, "", d)

View File

@ -1,6 +1,6 @@
# parse kernel ABI version out of <linux/version.h> # parse kernel ABI version out of <linux/version.h>
def get_kernelversion(p): def get_kernelversion(p):
import re, os import re
fn = p + '/include/linux/utsrelease.h' fn = p + '/include/linux/utsrelease.h'
if not os.path.isfile(fn): if not os.path.isfile(fn):
@ -30,7 +30,6 @@ def get_kernelmajorversion(p):
return None return None
def linux_module_packages(s, d): def linux_module_packages(s, d):
import bb, os.path
suffix = "" suffix = ""
return " ".join(map(lambda s: "kernel-module-%s%s" % (s.lower().replace('_', '-').replace('@', '+'), suffix), s.split())) return " ".join(map(lambda s: "kernel-module-%s%s" % (s.lower().replace('_', '-').replace('@', '+'), suffix), s.split()))

View File

@ -4,7 +4,6 @@ OPENMOKO_RELEASE ?= "OM-2007"
OPENMOKO_MIRROR ?= "svn://svn.openmoko.org/trunk" OPENMOKO_MIRROR ?= "svn://svn.openmoko.org/trunk"
def openmoko_base_get_subdir(d): def openmoko_base_get_subdir(d):
import bb
openmoko, section = bb.data.getVar('SECTION', d, 1).split("/") openmoko, section = bb.data.getVar('SECTION', d, 1).split("/")
if section == 'base' or section == 'libs': return "" if section == 'base' or section == 'libs': return ""
elif section in 'apps tools pim'.split(): return "applications" elif section in 'apps tools pim'.split(): return "applications"

View File

@ -5,12 +5,10 @@ OPENMOKO_RELEASE ?= "OM-2007.2"
OPENMOKO_MIRROR ?= "svn://svn.openmoko.org/trunk" OPENMOKO_MIRROR ?= "svn://svn.openmoko.org/trunk"
def openmoko_two_get_license(d): def openmoko_two_get_license(d):
import bb
openmoko, section = bb.data.getVar('SECTION', d, 1).split("/") openmoko, section = bb.data.getVar('SECTION', d, 1).split("/")
return "LGPL GPL".split()[section != "libs"] return "LGPL GPL".split()[section != "libs"]
def openmoko_two_get_subdir(d): def openmoko_two_get_subdir(d):
import bb
openmoko, section = bb.data.getVar('SECTION', d, 1).split("/") openmoko, section = bb.data.getVar('SECTION', d, 1).split("/")
if section == 'base': return "" if section == 'base': return ""
elif section == 'libs': return "libraries" elif section == 'libs': return "libraries"

View File

@ -29,7 +29,6 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
Used in .bb files to split up dynamically generated subpackages of a Used in .bb files to split up dynamically generated subpackages of a
given package, usually plugins or modules. given package, usually plugins or modules.
""" """
import os, os.path, bb
dvar = bb.data.getVar('PKGD', d, True) dvar = bb.data.getVar('PKGD', d, True)
@ -117,7 +116,6 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
PACKAGE_DEPENDS += "file-native" PACKAGE_DEPENDS += "file-native"
python () { python () {
import bb
if bb.data.getVar('PACKAGES', d, True) != '': if bb.data.getVar('PACKAGES', d, True) != '':
deps = bb.data.getVarFlag('do_package', 'depends', d) or "" deps = bb.data.getVarFlag('do_package', 'depends', d) or ""
for dep in (bb.data.getVar('PACKAGE_DEPENDS', d, True) or "").split(): for dep in (bb.data.getVar('PACKAGE_DEPENDS', d, True) or "").split():
@ -135,7 +133,7 @@ def runstrip(file, d):
# A working 'file' (one which works on the target architecture) # A working 'file' (one which works on the target architecture)
# is necessary for this stuff to work, hence the addition to do_package[depends] # is necessary for this stuff to work, hence the addition to do_package[depends]
import bb, os, commands, stat import commands, stat
pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True) pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
@ -192,8 +190,6 @@ def runstrip(file, d):
# #
def get_package_mapping (pkg, d): def get_package_mapping (pkg, d):
import bb, os
data = read_subpkgdata(pkg, d) data = read_subpkgdata(pkg, d)
key = "PKG_%s" % pkg key = "PKG_%s" % pkg
@ -203,8 +199,6 @@ def get_package_mapping (pkg, d):
return pkg return pkg
def runtime_mapping_rename (varname, d): def runtime_mapping_rename (varname, d):
import bb, os
#bb.note("%s before: %s" % (varname, bb.data.getVar(varname, d, True))) #bb.note("%s before: %s" % (varname, bb.data.getVar(varname, d, True)))
new_depends = [] new_depends = []
@ -226,8 +220,6 @@ def runtime_mapping_rename (varname, d):
# #
python package_do_split_locales() { python package_do_split_locales() {
import os
if (bb.data.getVar('PACKAGE_NO_LOCALE', d, True) == '1'): if (bb.data.getVar('PACKAGE_NO_LOCALE', d, True) == '1'):
bb.debug(1, "package requested not splitting locales") bb.debug(1, "package requested not splitting locales")
return return
@ -284,8 +276,6 @@ python package_do_split_locales() {
} }
python perform_packagecopy () { python perform_packagecopy () {
import os
dest = bb.data.getVar('D', d, True) dest = bb.data.getVar('D', d, True)
dvar = bb.data.getVar('PKGD', d, True) dvar = bb.data.getVar('PKGD', d, True)
@ -297,7 +287,7 @@ python perform_packagecopy () {
} }
python populate_packages () { python populate_packages () {
import os, glob, stat, errno, re import glob, stat, errno, re
workdir = bb.data.getVar('WORKDIR', d, True) workdir = bb.data.getVar('WORKDIR', d, True)
outdir = bb.data.getVar('DEPLOY_DIR', d, True) outdir = bb.data.getVar('DEPLOY_DIR', d, True)
@ -530,7 +520,7 @@ fi
SHLIBSDIR = "${STAGING_DIR_HOST}/shlibs" SHLIBSDIR = "${STAGING_DIR_HOST}/shlibs"
python package_do_shlibs() { python package_do_shlibs() {
import os, re, os.path import re
exclude_shlibs = bb.data.getVar('EXCLUDE_FROM_SHLIBS', d, 0) exclude_shlibs = bb.data.getVar('EXCLUDE_FROM_SHLIBS', d, 0)
if exclude_shlibs: if exclude_shlibs:
@ -746,7 +736,7 @@ python package_do_shlibs() {
} }
python package_do_pkgconfig () { python package_do_pkgconfig () {
import re, os import re
packages = bb.data.getVar('PACKAGES', d, True) packages = bb.data.getVar('PACKAGES', d, True)
workdir = bb.data.getVar('WORKDIR', d, True) workdir = bb.data.getVar('WORKDIR', d, True)

View File

@ -15,13 +15,11 @@ DPKG_ARCH_i686 ?= "i386"
DPKG_ARCH_pentium ?= "i386" DPKG_ARCH_pentium ?= "i386"
python package_deb_fn () { python package_deb_fn () {
from bb import data
bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d) bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
} }
addtask package_deb_install addtask package_deb_install
python do_package_deb_install () { python do_package_deb_install () {
import os, sys
pkg = bb.data.getVar('PKG', d, 1) pkg = bb.data.getVar('PKG', d, 1)
pkgfn = bb.data.getVar('PKGFN', d, 1) pkgfn = bb.data.getVar('PKGFN', d, 1)
rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1) rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
@ -37,6 +35,7 @@ python do_package_deb_install () {
os.makedirs(rootfs) os.makedirs(rootfs)
os.chdir(rootfs) os.chdir(rootfs)
except OSError: except OSError:
import sys
raise bb.build.FuncFailed(str(sys.exc_value)) raise bb.build.FuncFailed(str(sys.exc_value))
# update packages file # update packages file
@ -67,14 +66,13 @@ python do_package_deb_install () {
} }
python do_package_deb () { python do_package_deb () {
import sys, re, copy import re, copy
workdir = bb.data.getVar('WORKDIR', d, 1) workdir = bb.data.getVar('WORKDIR', d, 1)
if not workdir: if not workdir:
bb.error("WORKDIR not defined, unable to package") bb.error("WORKDIR not defined, unable to package")
return return
import os # path manipulations
outdir = bb.data.getVar('DEPLOY_DIR_DEB', d, 1) outdir = bb.data.getVar('DEPLOY_DIR_DEB', d, 1)
if not outdir: if not outdir:
bb.error("DEPLOY_DIR_DEB not defined, unable to package") bb.error("DEPLOY_DIR_DEB not defined, unable to package")
@ -135,8 +133,7 @@ python do_package_deb () {
except ValueError: except ValueError:
pass pass
if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1": if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
from bb import note bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PV', localdata, 1), bb.data.getVar('PR', localdata, 1)))
note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PV', localdata, 1), bb.data.getVar('PR', localdata, 1)))
bb.utils.unlockfile(lf) bb.utils.unlockfile(lf)
continue continue
@ -186,6 +183,7 @@ python do_package_deb () {
for (c, fs) in fields: for (c, fs) in fields:
ctrlfile.write(unicode(c % tuple(pullData(fs, localdata)))) ctrlfile.write(unicode(c % tuple(pullData(fs, localdata))))
except KeyError: except KeyError:
import sys
(type, value, traceback) = sys.exc_info() (type, value, traceback) = sys.exc_info()
bb.utils.unlockfile(lf) bb.utils.unlockfile(lf)
ctrlfile.close() ctrlfile.close()
@ -252,7 +250,6 @@ python do_package_deb () {
} }
python () { python () {
import bb
if bb.data.getVar('PACKAGES', d, True) != '': if bb.data.getVar('PACKAGES', d, True) != '':
deps = (bb.data.getVarFlag('do_package_write_deb', 'depends', d) or "").split() deps = (bb.data.getVarFlag('do_package_write_deb', 'depends', d) or "").split()
deps.append('dpkg-native:do_populate_staging') deps.append('dpkg-native:do_populate_staging')

View File

@ -6,12 +6,10 @@ IPKGCONF_TARGET = "${WORKDIR}/opkg.conf"
IPKGCONF_SDK = "${WORKDIR}/opkg-sdk.conf" IPKGCONF_SDK = "${WORKDIR}/opkg-sdk.conf"
python package_ipk_fn () { python package_ipk_fn () {
from bb import data
bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d) bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
} }
python package_ipk_install () { python package_ipk_install () {
import os, sys
pkg = bb.data.getVar('PKG', d, 1) pkg = bb.data.getVar('PKG', d, 1)
pkgfn = bb.data.getVar('PKGFN', d, 1) pkgfn = bb.data.getVar('PKGFN', d, 1)
rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1) rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
@ -25,6 +23,7 @@ python package_ipk_install () {
bb.mkdirhier(rootfs) bb.mkdirhier(rootfs)
os.chdir(rootfs) os.chdir(rootfs)
except OSError: except OSError:
import sys
(type, value, traceback) = sys.exc_info() (type, value, traceback) = sys.exc_info()
print value print value
raise bb.build.FuncFailed raise bb.build.FuncFailed
@ -126,14 +125,13 @@ package_generate_archlist () {
} }
python do_package_ipk () { python do_package_ipk () {
import sys, re, copy import re, copy
workdir = bb.data.getVar('WORKDIR', d, 1) workdir = bb.data.getVar('WORKDIR', d, 1)
if not workdir: if not workdir:
bb.error("WORKDIR not defined, unable to package") bb.error("WORKDIR not defined, unable to package")
return return
import os # path manipulations
outdir = bb.data.getVar('DEPLOY_DIR_IPK', d, 1) outdir = bb.data.getVar('DEPLOY_DIR_IPK', d, 1)
if not outdir: if not outdir:
bb.error("DEPLOY_DIR_IPK not defined, unable to package") bb.error("DEPLOY_DIR_IPK not defined, unable to package")
@ -192,8 +190,7 @@ python do_package_ipk () {
except ValueError: except ValueError:
pass pass
if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1": if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
from bb import note bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PV', localdata, 1), bb.data.getVar('PR', localdata, 1)))
note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PV', localdata, 1), bb.data.getVar('PR', localdata, 1)))
bb.utils.unlockfile(lf) bb.utils.unlockfile(lf)
continue continue
@ -234,6 +231,7 @@ python do_package_ipk () {
raise KeyError(f) raise KeyError(f)
ctrlfile.write(c % tuple(pullData(fs, localdata))) ctrlfile.write(c % tuple(pullData(fs, localdata)))
except KeyError: except KeyError:
import sys
(type, value, traceback) = sys.exc_info() (type, value, traceback) = sys.exc_info()
ctrlfile.close() ctrlfile.close()
bb.utils.unlockfile(lf) bb.utils.unlockfile(lf)
@ -302,7 +300,6 @@ python do_package_ipk () {
} }
python () { python () {
import bb
if bb.data.getVar('PACKAGES', d, True) != '': if bb.data.getVar('PACKAGES', d, True) != '':
deps = (bb.data.getVarFlag('do_package_write_ipk', 'depends', d) or "").split() deps = (bb.data.getVarFlag('do_package_write_ipk', 'depends', d) or "").split()
deps.append('opkg-utils-native:do_populate_staging') deps.append('opkg-utils-native:do_populate_staging')

View File

@ -12,9 +12,6 @@ RPMOPTS="--rcfile=${WORKDIR}/rpmrc --target ${TARGET_SYS}"
RPM="rpm ${RPMOPTS}" RPM="rpm ${RPMOPTS}"
python write_specfile() { python write_specfile() {
from bb import data, build
import sys
version = bb.data.getVar('PV', d, 1) version = bb.data.getVar('PV', d, 1)
version = version.replace('-', '+') version = version.replace('-', '+')
bb.data.setVar('RPMPV', version, d) bb.data.setVar('RPMPV', version, d)
@ -55,8 +52,7 @@ python write_specfile() {
pass pass
if not files and bb.data.getVar('ALLOW_EMPTY', d) != "1": if not files and bb.data.getVar('ALLOW_EMPTY', d) != "1":
from bb import note bb.note("Not creating empty archive for %s-%s-%s" % (bb.data.getVar('PKG',d, 1), bb.data.getVar('PV', d, 1), bb.data.getVar('PR', d, 1)))
note("Not creating empty archive for %s-%s-%s" % (bb.data.getVar('PKG',d, 1), bb.data.getVar('PV', d, 1), bb.data.getVar('PR', d, 1)))
return return
# output .spec using this metadata store # output .spec using this metadata store
@ -159,7 +155,6 @@ python do_package_rpm () {
bb.error("WORKDIR not defined, unable to package") bb.error("WORKDIR not defined, unable to package")
return return
import os # path manipulations
outdir = bb.data.getVar('DEPLOY_DIR_RPM', d, 1) outdir = bb.data.getVar('DEPLOY_DIR_RPM', d, 1)
if not outdir: if not outdir:
bb.error("DEPLOY_DIR_RPM not defined, unable to package") bb.error("DEPLOY_DIR_RPM not defined, unable to package")
@ -213,7 +208,6 @@ python do_package_rpm () {
} }
python () { python () {
import bb
if bb.data.getVar('PACKAGES', d, True) != '': if bb.data.getVar('PACKAGES', d, True) != '':
deps = (bb.data.getVarFlag('do_package_write_rpm', 'depends', d) or "").split() deps = (bb.data.getVarFlag('do_package_write_rpm', 'depends', d) or "").split()
deps.append('rpm-native:do_populate_staging') deps.append('rpm-native:do_populate_staging')

View File

@ -3,15 +3,12 @@ inherit package
IMAGE_PKGTYPE ?= "tar" IMAGE_PKGTYPE ?= "tar"
python package_tar_fn () { python package_tar_fn () {
import os
from bb import data
fn = os.path.join(bb.data.getVar('DEPLOY_DIR_TAR', d), "%s-%s-%s.tar.gz" % (bb.data.getVar('PKG', d), bb.data.getVar('PV', d), bb.data.getVar('PR', d))) fn = os.path.join(bb.data.getVar('DEPLOY_DIR_TAR', d), "%s-%s-%s.tar.gz" % (bb.data.getVar('PKG', d), bb.data.getVar('PV', d), bb.data.getVar('PR', d)))
fn = bb.data.expand(fn, d) fn = bb.data.expand(fn, d)
bb.data.setVar('PKGFN', fn, d) bb.data.setVar('PKGFN', fn, d)
} }
python package_tar_install () { python package_tar_install () {
import os, sys
pkg = bb.data.getVar('PKG', d, 1) pkg = bb.data.getVar('PKG', d, 1)
pkgfn = bb.data.getVar('PKGFN', d, 1) pkgfn = bb.data.getVar('PKGFN', d, 1)
rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1) rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
@ -23,6 +20,7 @@ python package_tar_install () {
bb.mkdirhier(rootfs) bb.mkdirhier(rootfs)
os.chdir(rootfs) os.chdir(rootfs)
except OSError: except OSError:
import sys
(type, value, traceback) = sys.exc_info() (type, value, traceback) = sys.exc_info()
print value print value
raise bb.build.FuncFailed raise bb.build.FuncFailed
@ -42,7 +40,6 @@ python do_package_tar () {
bb.error("WORKDIR not defined, unable to package") bb.error("WORKDIR not defined, unable to package")
return return
import os # path manipulations
outdir = bb.data.getVar('DEPLOY_DIR_TAR', d, 1) outdir = bb.data.getVar('DEPLOY_DIR_TAR', d, 1)
if not outdir: if not outdir:
bb.error("DEPLOY_DIR_TAR not defined, unable to package") bb.error("DEPLOY_DIR_TAR not defined, unable to package")
@ -94,7 +91,6 @@ python do_package_tar () {
} }
python () { python () {
import bb
if bb.data.getVar('PACKAGES', d, True) != '': if bb.data.getVar('PACKAGES', d, True) != '':
deps = (bb.data.getVarFlag('do_package_write_tar', 'depends', d) or "").split() deps = (bb.data.getVarFlag('do_package_write_tar', 'depends', d) or "").split()
deps.append('tar-native:do_populate_staging') deps.append('tar-native:do_populate_staging')

View File

@ -27,7 +27,6 @@ PSTAGE_NATIVEDEPENDS = "\
BB_STAMP_WHITELIST = "${PSTAGE_NATIVEDEPENDS}" BB_STAMP_WHITELIST = "${PSTAGE_NATIVEDEPENDS}"
python () { python () {
import bb
pstage_allowed = True pstage_allowed = True
# These classes encode staging paths into the binary data so can only be # These classes encode staging paths into the binary data so can only be
@ -81,8 +80,6 @@ PSTAGE_LIST_CMD = "${PSTAGE_PKGMANAGER} -f ${PSTAGE_MACHCONFIG} -o ${TMP
PSTAGE_TMPDIR_STAGE = "${WORKDIR}/staging-pkg" PSTAGE_TMPDIR_STAGE = "${WORKDIR}/staging-pkg"
def pstage_manualclean(srcname, destvarname, d): def pstage_manualclean(srcname, destvarname, d):
import os, bb
src = os.path.join(bb.data.getVar('PSTAGE_TMPDIR_STAGE', d, True), srcname) src = os.path.join(bb.data.getVar('PSTAGE_TMPDIR_STAGE', d, True), srcname)
dest = bb.data.getVar(destvarname, d, True) dest = bb.data.getVar(destvarname, d, True)
@ -95,7 +92,6 @@ def pstage_manualclean(srcname, destvarname, d):
os.system("rm %s 2> /dev/null" % filepath) os.system("rm %s 2> /dev/null" % filepath)
def pstage_set_pkgmanager(d): def pstage_set_pkgmanager(d):
import bb
path = bb.data.getVar("PATH", d, 1) path = bb.data.getVar("PATH", d, 1)
pkgmanager = bb.which(path, 'opkg-cl') pkgmanager = bb.which(path, 'opkg-cl')
if pkgmanager == "": if pkgmanager == "":
@ -105,8 +101,6 @@ def pstage_set_pkgmanager(d):
def pstage_cleanpackage(pkgname, d): def pstage_cleanpackage(pkgname, d):
import os, bb
path = bb.data.getVar("PATH", d, 1) path = bb.data.getVar("PATH", d, 1)
pstage_set_pkgmanager(d) pstage_set_pkgmanager(d)
list_cmd = bb.data.getVar("PSTAGE_LIST_CMD", d, True) list_cmd = bb.data.getVar("PSTAGE_LIST_CMD", d, True)
@ -168,8 +162,6 @@ PSTAGE_TASKS_COVERED = "fetch unpack munge patch configure qa_configure rig_loca
SCENEFUNCS += "packagestage_scenefunc" SCENEFUNCS += "packagestage_scenefunc"
python packagestage_scenefunc () { python packagestage_scenefunc () {
import os
if bb.data.getVar("PSTAGING_ACTIVE", d, 1) == "0": if bb.data.getVar("PSTAGING_ACTIVE", d, 1) == "0":
return return
@ -249,10 +241,7 @@ packagestage_scenefunc[dirs] = "${STAGING_DIR}"
addhandler packagedstage_stampfixing_eventhandler addhandler packagedstage_stampfixing_eventhandler
python packagedstage_stampfixing_eventhandler() { python packagedstage_stampfixing_eventhandler() {
from bb.event import getName if bb.event.getName(e) == "StampUpdate":
import os
if getName(e) == "StampUpdate":
taskscovered = bb.data.getVar("PSTAGE_TASKS_COVERED", e.data, 1).split() taskscovered = bb.data.getVar("PSTAGE_TASKS_COVERED", e.data, 1).split()
for (fn, task) in e.targets: for (fn, task) in e.targets:
# strip off 'do_' # strip off 'do_'

View File

@ -1,5 +1,4 @@
def packaged(pkg, d): def packaged(pkg, d):
import os, bb
return os.access(get_subpkgedata_fn(pkg, d) + '.packaged', os.R_OK) return os.access(get_subpkgedata_fn(pkg, d) + '.packaged', os.R_OK)
def read_pkgdatafile(fn): def read_pkgdatafile(fn):
@ -10,7 +9,6 @@ def read_pkgdatafile(fn):
c = codecs.getdecoder("string_escape") c = codecs.getdecoder("string_escape")
return c(str)[0] return c(str)[0]
import os
if os.access(fn, os.R_OK): if os.access(fn, os.R_OK):
import re import re
f = file(fn, 'r') f = file(fn, 'r')
@ -25,7 +23,6 @@ def read_pkgdatafile(fn):
return pkgdata return pkgdata
def get_subpkgedata_fn(pkg, d): def get_subpkgedata_fn(pkg, d):
import bb, os
archs = bb.data.expand("${PACKAGE_ARCHS}", d).split(" ") archs = bb.data.expand("${PACKAGE_ARCHS}", d).split(" ")
archs.reverse() archs.reverse()
pkgdata = bb.data.expand('${TMPDIR}/pkgdata/', d) pkgdata = bb.data.expand('${TMPDIR}/pkgdata/', d)
@ -37,25 +34,20 @@ def get_subpkgedata_fn(pkg, d):
return bb.data.expand('${PKGDATA_DIR}/runtime/%s' % pkg, d) return bb.data.expand('${PKGDATA_DIR}/runtime/%s' % pkg, d)
def has_subpkgdata(pkg, d): def has_subpkgdata(pkg, d):
import bb, os
return os.access(get_subpkgedata_fn(pkg, d), os.R_OK) return os.access(get_subpkgedata_fn(pkg, d), os.R_OK)
def read_subpkgdata(pkg, d): def read_subpkgdata(pkg, d):
import bb
return read_pkgdatafile(get_subpkgedata_fn(pkg, d)) return read_pkgdatafile(get_subpkgedata_fn(pkg, d))
def has_pkgdata(pn, d): def has_pkgdata(pn, d):
import bb, os
fn = bb.data.expand('${PKGDATA_DIR}/%s' % pn, d) fn = bb.data.expand('${PKGDATA_DIR}/%s' % pn, d)
return os.access(fn, os.R_OK) return os.access(fn, os.R_OK)
def read_pkgdata(pn, d): def read_pkgdata(pn, d):
import bb
fn = bb.data.expand('${PKGDATA_DIR}/%s' % pn, d) fn = bb.data.expand('${PKGDATA_DIR}/%s' % pn, d)
return read_pkgdatafile(fn) return read_pkgdatafile(fn)
python read_subpackage_metadata () { python read_subpackage_metadata () {
import bb
data = read_pkgdata(bb.data.getVar('PN', d, 1), d) data = read_pkgdata(bb.data.getVar('PN', d, 1), d)
for key in data.keys(): for key in data.keys():
@ -72,7 +64,6 @@ python read_subpackage_metadata () {
# Collapse FOO_pkg variables into FOO # Collapse FOO_pkg variables into FOO
# #
def read_subpkgdata_dict(pkg, d): def read_subpkgdata_dict(pkg, d):
import bb
ret = {} ret = {}
subd = read_pkgdatafile(get_subpkgedata_fn(pkg, d)) subd = read_pkgdatafile(get_subpkgedata_fn(pkg, d))
for var in subd: for var in subd:

View File

@ -61,8 +61,6 @@ python emit_pkghistory() {
def check_pkghistory(pkg, pe, pv, pr, lastversion): def check_pkghistory(pkg, pe, pv, pr, lastversion):
import bb
(last_pe, last_pv, last_pr) = lastversion (last_pe, last_pv, last_pr) = lastversion
bb.debug(2, "Checking package history") bb.debug(2, "Checking package history")
@ -72,7 +70,6 @@ def check_pkghistory(pkg, pe, pv, pr, lastversion):
def write_pkghistory(pkg, pe, pv, pr, d): def write_pkghistory(pkg, pe, pv, pr, d):
import bb, os
bb.debug(2, "Writing package history") bb.debug(2, "Writing package history")
pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True) pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
@ -82,8 +79,6 @@ def write_pkghistory(pkg, pe, pv, pr, d):
os.makedirs(verpath) os.makedirs(verpath)
def write_latestlink(pkg, pe, pv, pr, d): def write_latestlink(pkg, pe, pv, pr, d):
import bb, os
pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True) pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
def rm_link(path): def rm_link(path):

View File

@ -4,8 +4,6 @@
QUILTRCFILE ?= "${STAGING_BINDIR_NATIVE}/quiltrc" QUILTRCFILE ?= "${STAGING_BINDIR_NATIVE}/quiltrc"
def patch_init(d): def patch_init(d):
import os, sys
class NotFoundError(Exception): class NotFoundError(Exception):
def __init__(self, path): def __init__(self, path):
self.path = path self.path = path
@ -13,8 +11,6 @@ def patch_init(d):
return "Error: %s not found." % self.path return "Error: %s not found." % self.path
def md5sum(fname): def md5sum(fname):
import sys
# when we move to Python 2.5 as minimal supported # when we move to Python 2.5 as minimal supported
# we can kill that try/except as hashlib is 2.5+ # we can kill that try/except as hashlib is 2.5+
try: try:
@ -76,8 +72,6 @@ def patch_init(d):
def __str__(self): def __str__(self):
return "Patch Error: %s" % self.msg return "Patch Error: %s" % self.msg
import bb, bb.data, bb.fetch
class PatchSet(object): class PatchSet(object):
defaults = { defaults = {
"strippath": 1 "strippath": 1
@ -251,6 +245,7 @@ def patch_init(d):
try: try:
output = runcmd(["quilt", "applied"], self.dir) output = runcmd(["quilt", "applied"], self.dir)
except CmdError: except CmdError:
import sys
if sys.exc_value.output.strip() == "No patches applied": if sys.exc_value.output.strip() == "No patches applied":
return return
else: else:
@ -364,6 +359,7 @@ def patch_init(d):
try: try:
self.patchset.Push() self.patchset.Push()
except Exception: except Exception:
import sys
os.chdir(olddir) os.chdir(olddir)
raise sys.exc_value raise sys.exc_value
@ -458,9 +454,6 @@ PATCHDEPENDENCY = "${PATCHTOOL}-native:do_populate_staging"
do_patch[depends] = "${PATCHDEPENDENCY}" do_patch[depends] = "${PATCHDEPENDENCY}"
python patch_do_patch() { python patch_do_patch() {
import re
import bb.fetch
patch_init(d) patch_init(d)
src_uri = (bb.data.getVar('SRC_URI', d, 1) or '').split() src_uri = (bb.data.getVar('SRC_URI', d, 1) or '').split()

View File

@ -159,7 +159,7 @@ ipk_insert_feed_uris () {
} }
python () { python () {
import bb
if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True): if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True):
flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d) flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d)
flags = flags.replace("do_package_write_ipk", "") flags = flags.replace("do_package_write_ipk", "")

View File

@ -234,7 +234,6 @@ install_all_locales() {
} }
python () { python () {
import bb
if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True): if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True):
flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d) flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d)
flags = flags.replace("do_package_write_rpm", "") flags = flags.replace("do_package_write_rpm", "")

View File

@ -3,7 +3,6 @@
# #
def raise_sanity_error(msg): def raise_sanity_error(msg):
import bb
bb.fatal(""" Poky's config sanity checker detected a potential misconfiguration. bb.fatal(""" Poky's config sanity checker detected a potential misconfiguration.
Either fix the cause of this error or at your own risk disable the checker (see sanity.conf). Either fix the cause of this error or at your own risk disable the checker (see sanity.conf).
Following is the list of potential problems / advisories: Following is the list of potential problems / advisories:
@ -11,8 +10,6 @@ def raise_sanity_error(msg):
%s""" % msg) %s""" % msg)
def check_conf_exists(fn, data): def check_conf_exists(fn, data):
import bb, os
bbpath = [] bbpath = []
fn = bb.data.expand(fn, data) fn = bb.data.expand(fn, data)
vbbpath = bb.data.getVar("BBPATH", data) vbbpath = bb.data.getVar("BBPATH", data)
@ -26,12 +23,12 @@ def check_conf_exists(fn, data):
def check_sanity(e): def check_sanity(e):
from bb import note, error, data, __version__ from bb import note, error, data, __version__
from bb.event import Handled, NotHandled, getName
try: try:
from distutils.version import LooseVersion from distutils.version import LooseVersion
except ImportError: except ImportError:
def LooseVersion(v): print "WARNING: sanity.bbclass can't compare versions without python-distutils"; return 1 def LooseVersion(v): print "WARNING: sanity.bbclass can't compare versions without python-distutils"; return 1
import os, commands import commands
# Check the bitbake version meets minimum requirements # Check the bitbake version meets minimum requirements
minversion = data.getVar('BB_MIN_VERSION', e.data , True) minversion = data.getVar('BB_MIN_VERSION', e.data , True)
@ -163,10 +160,8 @@ def check_sanity(e):
addhandler check_sanity_eventhandler addhandler check_sanity_eventhandler
python check_sanity_eventhandler() { python check_sanity_eventhandler() {
from bb import note, error, data, __version__ from bb.event import Handled, NotHandled
from bb.event import getName if bb.event.getName(e) == "ConfigParsed":
if getName(e) == "ConfigParsed":
check_sanity(e) check_sanity(e)
return NotHandled return NotHandled

View File

@ -16,8 +16,6 @@
# If 'what' doesn't exist then an empty value is returned # If 'what' doesn't exist then an empty value is returned
# #
def get_siteinfo_list(d): def get_siteinfo_list(d):
import bb
target = bb.data.getVar('HOST_ARCH', d, 1) + "-" + bb.data.getVar('HOST_OS', d, 1) target = bb.data.getVar('HOST_ARCH', d, 1) + "-" + bb.data.getVar('HOST_OS', d, 1)
targetinfo = {\ targetinfo = {\
@ -74,8 +72,6 @@ def get_siteinfo_list(d):
# 2) ${FILE_DIRNAME}/site-${PV} - app version specific # 2) ${FILE_DIRNAME}/site-${PV} - app version specific
# #
def siteinfo_get_files(d): def siteinfo_get_files(d):
import bb, os
sitefiles = "" sitefiles = ""
# Determine which site files to look for # Determine which site files to look for

View File

@ -5,8 +5,6 @@ EXCLUDE_FROM ?= ".pc autom4te.cache"
DISTRO ?= "openembedded" DISTRO ?= "openembedded"
def get_src_tree(d): def get_src_tree(d):
import bb
import os, os.path
workdir = bb.data.getVar('WORKDIR', d, 1) workdir = bb.data.getVar('WORKDIR', d, 1)
if not workdir: if not workdir:
@ -56,8 +54,6 @@ sourcepkg_do_archive_bb() {
} }
python sourcepkg_do_dumpdata() { python sourcepkg_do_dumpdata() {
import os
import os.path
workdir = bb.data.getVar('WORKDIR', d, 1) workdir = bb.data.getVar('WORKDIR', d, 1)
distro = bb.data.getVar('DISTRO', d, 1) distro = bb.data.getVar('DISTRO', d, 1)

View File

@ -51,8 +51,7 @@ def tinder_format_http_post(d,status,log):
for the tinderbox to be happy. for the tinderbox to be happy.
""" """
from bb import data, build import random
import os,random
# the variables we will need to send on this form post # the variables we will need to send on this form post
variables = { variables = {
@ -125,7 +124,6 @@ def tinder_build_start(d):
report = report[report.find(search)+len(search):] report = report[report.find(search)+len(search):]
report = report[0:report.find("'")] report = report[0:report.find("'")]
import bb
bb.note("Machine ID assigned by tinderbox: %s" % report ) bb.note("Machine ID assigned by tinderbox: %s" % report )
# now we will need to save the machine number # now we will need to save the machine number
@ -165,7 +163,6 @@ def tinder_print_info(d):
""" """
from bb import data from bb import data
import os
# get the local vars # get the local vars
time = tinder_time_string() time = tinder_time_string()
@ -216,7 +213,6 @@ def tinder_print_env():
Print the environment variables of this build Print the environment variables of this build
""" """
from bb import data from bb import data
import os
time_start = tinder_time_string() time_start = tinder_time_string()
time_end = tinder_time_string() time_end = tinder_time_string()
@ -278,7 +274,7 @@ def tinder_do_tinder_report(event):
""" """
from bb.event import getName from bb.event import getName
from bb import data, mkdirhier, build from bb import data, mkdirhier, build
import os, glob import glob
# variables # variables
name = getName(event) name = getName(event)

View File

@ -11,7 +11,6 @@ update-alternatives --remove ${ALTERNATIVE_NAME} ${ALTERNATIVE_PATH}
} }
def update_alternatives_after_parse(d): def update_alternatives_after_parse(d):
import bb
if bb.data.getVar('ALTERNATIVE_NAME', d) == None: if bb.data.getVar('ALTERNATIVE_NAME', d) == None:
raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_NAME" % bb.data.getVar('FILE', d) raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_NAME" % bb.data.getVar('FILE', d)
if bb.data.getVar('ALTERNATIVE_PATH', d) == None: if bb.data.getVar('ALTERNATIVE_PATH', d) == None:

View File

@ -26,7 +26,6 @@ update-rc.d $D ${INITSCRIPT_NAME} remove
def update_rc_after_parse(d): def update_rc_after_parse(d):
import bb
if bb.data.getVar('INITSCRIPT_PACKAGES', d) == None: if bb.data.getVar('INITSCRIPT_PACKAGES', d) == None:
if bb.data.getVar('INITSCRIPT_NAME', d) == None: if bb.data.getVar('INITSCRIPT_NAME', d) == None:
raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_NAME" % bb.data.getVar('FILE', d) raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_NAME" % bb.data.getVar('FILE', d)