From 90a211eb560d7c984a444f9697ff1d06e4bca838 Mon Sep 17 00:00:00 2001 From: Holger Hans Peter Freyther Date: Wed, 30 May 2012 11:20:26 +0200 Subject: [PATCH] meta-sysmocom-bsp: Backport the archives --- classes/archive-configured-source.bbclass | 14 + classes/archive-original-source.bbclass | 14 + classes/archive-patched-source.bbclass | 14 + classes/archiver.bbclass | 450 ++++++++++++++++++++++ 4 files changed, 492 insertions(+) create mode 100644 classes/archive-configured-source.bbclass create mode 100644 classes/archive-original-source.bbclass create mode 100644 classes/archive-patched-source.bbclass create mode 100644 classes/archiver.bbclass diff --git a/classes/archive-configured-source.bbclass b/classes/archive-configured-source.bbclass new file mode 100644 index 0000000..1a609b3 --- /dev/null +++ b/classes/archive-configured-source.bbclass @@ -0,0 +1,14 @@ +# This file is for getting archiving packages with configured sources(archive 's' after configure stage),logs(archive 'temp' after package_write_rpm),dump data +# and creating diff file(get all environment variables and functions in building and mapping all content in 's' including patches to xxx.diff.gz. +# All archived packages will be deployed in ${DEPLOY_DIR}/sources + +inherit archiver + +# Get archiving package with configured sources including patches +do_configure[postfuncs] += "do_archive_configured_sources " + +# Get archiving package with temp(logs) and scripts(.bb and inc files) +do_package_write_rpm[prefuncs] += "do_archive_scripts_logs " + +# Get dump date and create diff file +do_package_write_rpm[postfuncs] += "do_dumpdata_create_diff_gz " diff --git a/classes/archive-original-source.bbclass b/classes/archive-original-source.bbclass new file mode 100644 index 0000000..b085533 --- /dev/null +++ b/classes/archive-original-source.bbclass @@ -0,0 +1,14 @@ +# This file is for getting archiving packages with original sources(archive 's' after unpack stage),patches,logs(archive 'temp' after package_write_rpm),dump data and +# creating diff file(get all environment variables and functions in building and mapping all content in 's' including patches to xxx.diff.gz. +# All archived packages will be deployed in ${DEPLOY_DIR}/sources + +inherit archiver + +# Get original sources archiving package with patches +do_unpack[postfuncs] += "do_archive_original_sources_patches " + +# Get archiving package with temp(logs) and scripts(.bb and inc files) +do_package_write_rpm[prefuncs] += "do_archive_scripts_logs " + +# Get dump date and create diff file +do_package_write_rpm[postfuncs] += "do_dumpdata_create_diff_gz " diff --git a/classes/archive-patched-source.bbclass b/classes/archive-patched-source.bbclass new file mode 100644 index 0000000..a6d368f --- /dev/null +++ b/classes/archive-patched-source.bbclass @@ -0,0 +1,14 @@ +# This file is for getting archiving packages with patched sources(archive 's' before do_patch stage),logs(archive 'temp' after package_write_rpm),dump data and +# creating diff file(get all environment variables and functions in building and mapping all content in 's' including patches to xxx.diff.gz. +# All archived packages will be deployed in ${DEPLOY_DIR}/sources + +inherit archiver + +# Get archiving package with patched sources including patches +do_patch[postfuncs] += "do_archive_patched_sources " + +# Get archiving package with logs(temp) and scripts(.bb and .inc files) +do_package_write_rpm[prefuncs] += "do_archive_scripts_logs " + +# Get dump date and create diff file +do_package_write_rpm[postfuncs] += "do_dumpdata_create_diff_gz " diff --git a/classes/archiver.bbclass b/classes/archiver.bbclass new file mode 100644 index 0000000..59b58f4 --- /dev/null +++ b/classes/archiver.bbclass @@ -0,0 +1,450 @@ +# This file is used for archiving sources ,patches,and logs to tarball. +# It also output building environment to xxx.dump.data and create xxx.diff.gz to record +# all content in ${S} to a diff file. + +ARCHIVE_EXCLUDE_FROM ?= ".pc autom4te.cache" +ARCHIVE_TYPE ?= "TAR SRPM" +DISTRO ?= "poky" +PATCHES_ARCHIVE_WITH_SERIES = 'TRUE' + +def get_bb_inc(d): + '''create a directory "script-logs" including .bb and .inc file in ${WORKDIR}''' + import re + import os + import shutil + + bbinc = [] + pat=re.compile('require\s*([^\s]*\.*)(.*)') + work_dir = d.getVar('WORKDIR', True) + bbfile = d.getVar('FILE', True) + bbdir = os.path.dirname(bbfile) + script_logs = os.path.join(work_dir,'script-logs') + bb_inc = os.path.join(script_logs,'bb_inc') + bb.mkdirhier(script_logs) + bb.mkdirhier(bb_inc) + + def find_file(dir,file): + for root, dirs, files in os.walk(dir): + if file in files: + return os.path.join(root,file) + + def get_inc (file): + f = open(file,'r') + for line in f.readlines(): + if 'require' not in line: + bbinc.append(file) + else: + try: + incfile = pat.match(line).group(1) + incfile = bb.data.expand(os.path.basename(incfile),d) + abs_incfile = find_file(bbdir,incfile) + if abs_incfile: + bbinc.append(abs_incfile) + get_inc(abs_incfile) + except AttributeError: + pass + get_inc(bbfile) + bbinc = list(set(bbinc)) + for bbincfile in bbinc: + shutil.copy(bbincfile,bb_inc) + + try: + bb.mkdirhier(os.path.join(script_logs,'temp')) + oe.path.copytree(os.path.join(work_dir,'temp'), os.path.join(script_logs,'temp')) + except (IOError,AttributeError): + pass + return script_logs + +def get_series(d): + '''copy patches and series file to a pointed directory which will be archived to tarball in ${WORKDIR}''' + import shutil + + src_patches=[] + pf = d.getVar('PF', True) + work_dir = d.getVar('WORKDIR', True) + s = d.getVar('S',True) + dest = os.path.join(work_dir, pf + '-series') + shutil.rmtree(dest, ignore_errors=True) + bb.mkdirhier(dest) + + src_uri = d.getVar('SRC_URI', True).split() + fetch = bb.fetch2.Fetch(src_uri, d) + locals = (fetch.localpath(url) for url in fetch.urls) + for local in locals: + src_patches.append(local) + if not cmp(work_dir,s): + tmp_list = src_patches + else: + tmp_list = src_patches[1:] + + for patch in tmp_list: + try: + shutil.copy(patch,dest) + except IOError: + if os.path.isdir(patch): + bb.mkdirhier(os.path.join(dest,patch)) + oe.path.copytree(patch, os.path.join(dest,patch)) + return dest + +def get_applying_patches(d): + """only copy applying patches to a pointed directory which will be archived to tarball""" + import os + import shutil + + + pf = d.getVar('PF', True) + work_dir = d.getVar('WORKDIR', True) + dest = os.path.join(work_dir, pf + '-patches') + shutil.rmtree(dest, ignore_errors=True) + bb.mkdirhier(dest) + + + patches = src_patches(d) + for patch in patches: + _, _, local, _, _, parm = bb.decodeurl(patch) + if local: + shutil.copy(local,dest) + return dest + +def not_tarball(d): + '''packages including key words 'work-shared','native', 'task-' will be passed''' + import os + + workdir = d.getVar('WORKDIR',True) + s = d.getVar('S',True) + if 'work-shared' in s or 'task-' in workdir or 'native' in workdir: + return True + else: + return False + +def get_source_from_downloads(d,stage_name): + '''copy tarball of $P to $WORKDIR when this tarball exists in $DL_DIR''' + if stage_name in 'patched' 'configured': + return + pf = d.getVar('PF', True) + dl_dir = d.getVar('DL_DIR',True) + try: + source = os.path.join(dl_dir,os.path.basename(d.getVar('SRC_URI', True).split()[0])) + if os.path.exists(source) and not os.path.isdir(source): + return source + except (IndexError, OSError): + pass + return '' + +def do_tarball(workdir,srcdir,tarname): + '''tar "srcdir" under "workdir" to "tarname"''' + import tarfile + + sav_dir = os.getcwd() + os.chdir(workdir) + if (len(os.listdir(srcdir))) != 0: + tar = tarfile.open(tarname, "w:gz") + tar.add(srcdir) + tar.close() + else: + tarname = '' + os.chdir(sav_dir) + return tarname + +def archive_sources_from_directory(d,stage_name): + '''archive sources codes tree to tarball when tarball of $P doesn't exist in $DL_DIR''' + import shutil + + s = d.getVar('S',True) + work_dir=d.getVar('WORKDIR', True) + PF = d.getVar('PF',True) + tarname = PF + '-' + stage_name + ".tar.gz" + + if os.path.exists(s) and work_dir in s: + try: + source_dir = os.path.join(work_dir,[ i for i in s.replace(work_dir,'').split('/') if i][0]) + except IndexError: + if not cmp(s,work_dir): + return '' + else: + return '' + source = os.path.basename(source_dir) + return do_tarball(work_dir,source,tarname) + +def archive_sources(d,stage_name): + '''copy tarball from $DL_DIR to $WORKDIR if have tarball, archive source codes tree in $WORKDIR if $P is directory instead of tarball''' + import shutil + work_dir = d.getVar('WORKDIR',True) + file = get_source_from_downloads(d,stage_name) + if file: + shutil.copy(file,work_dir) + file = os.path.basename(file) + else: + file = archive_sources_from_directory(d,stage_name) + return file + + +def archive_patches(d,patchdir,series): + '''archive patches to tarball and also include series files if 'series' is True''' + import shutil + + s = d.getVar('S',True) + work_dir = d.getVar('WORKDIR', True) + patch_dir = os.path.basename(patchdir) + tarname = patch_dir + ".tar.gz" + if series == 'all' and os.path.exists(os.path.join(s,'patches/series')): + shutil.copy(os.path.join(s,'patches/series'),patchdir) + tarname = do_tarball(work_dir,patch_dir,tarname) + shutil.rmtree(patchdir, ignore_errors=True) + return tarname + +def select_archive_patches(d,option): + '''select to archive all patches including non-applying and series or applying patches ''' + if option == "all": + patchdir = get_series(d) + elif option == "applying": + patchdir = get_applying_patches(d) + try: + os.rmdir(patchdir) + except OSError: + tarpatch = archive_patches(d,patchdir,option) + return tarpatch + return + +def archive_logs(d,logdir,bbinc=False): + '''archive logs in temp to tarball and .bb and .inc files if bbinc is True ''' + import shutil + + pf = d.getVar('PF',True) + work_dir = d.getVar('WORKDIR',True) + log_dir = os.path.basename(logdir) + tarname = pf + '-' + log_dir + ".tar.gz" + tarname = do_tarball(work_dir,log_dir,tarname) + if bbinc: + shutil.rmtree(logdir, ignore_errors=True) + return tarname + +def get_licenses(d): + '''get licenses for running .bb file''' + licenses = d.getVar('LICENSE', 1).replace('&', '|') + licenses = licenses.replace('(', '').replace(')', '') + clean_licenses = "" + for x in licenses.split(): + if x.strip() == '' or x == 'CLOSED': + continue + if x != "|": + clean_licenses += x + if '|' in clean_licenses: + clean_licenses = clean_licenses.replace('|','') + return clean_licenses + + +def move_tarball_deploy(d,tarball_list): + '''move tarball in location to ${DEPLOY_DIR}/sources''' + import shutil + + if tarball_list is []: + return + target_sys = d.getVar('TARGET_SYS', True) + pf = d.getVar('PF', True) + licenses = get_licenses(d) + work_dir = d.getVar('WORKDIR',True) + tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + if not os.path.exists(tar_sources): + bb.mkdirhier(tar_sources) + for source in tarball_list: + if source: + if os.path.exists(os.path.join(tar_sources, source)): + os.remove(os.path.join(tar_sources,source)) + shutil.move(os.path.join(work_dir,source),tar_sources) + +def check_archiving_type(d): + '''check the type for archiving package('tar' or 'srpm')''' + try: + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in d.getVar('ARCHIVE_TYPE', True).split(): + raise AttributeError + except AttributeError: + bb.fatal("\"SOURCE_ARCHIVE_PACKAGE_TYPE\" is \'tar\' or \'srpm\', no other types") + +def store_package(d,package_name): + '''store tarbablls name to file "tar-package"''' + try: + f = open(os.path.join(d.getVar('WORKDIR',True),'tar-package'),'a') + f.write(package_name + ' ') + f.close() + except IOError: + pass + +def get_package(d): + '''get tarballs name from "tar-package"''' + work_dir = (d.getVar('WORKDIR', True)) + tarpackage = os.path.join(work_dir,'tar-package') + try: + f = open(tarpackage,'r') + line = list(set(f.readline().replace('\n','').split())) + except IOError: + pass + f.close() + return line + + +def archive_sources_patches(d,stage_name): + '''archive sources and patches to tarball. stage_name will append strings ${stage_name} to ${PR} as middle name. for example, zlib-1.4.6-prepatch(stage_name).tar.gz ''' + import shutil + + check_archiving_type(d) + if not_tarball(d): + return + + source_tar_name = archive_sources(d,stage_name) + if stage_name == "prepatch": + if d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'TRUE': + patch_tar_name = select_archive_patches(d,"all") + elif d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'FALSE': + patch_tar_name = select_archive_patches(d,"applying") + else: + bb.fatal("Please define 'PATCHES_ARCHIVE_WITH_SERIES' is strings 'True' or 'False' ") + else: + patch_tar_name = '' + + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[source_tar_name,patch_tar_name]) + else: + tarpackage = os.path.join(d.getVar('WORKDIR', True),'tar-package') + if os.path.exists(tarpackage): + os.remove(tarpackage) + for package in os.path.basename(source_tar_name), patch_tar_name: + if package: + store_package(d,str(package) + ' ') + +def archive_scripts_logs(d): + '''archive scripts and logs. scripts include .bb and .inc files and logs include stuff in "temp".''' + + work_dir = d.getVar('WORKDIR', True) + temp_dir = os.path.join(work_dir,'temp') + source_archive_log_with_scripts = d.getVar('SOURCE_ARCHIVE_LOG_WITH_SCRIPTS', True) + if source_archive_log_with_scripts == 'logs_with_scripts': + logdir = get_bb_inc(d) + tarlog = archive_logs(d,logdir,True) + elif source_archive_log_with_scripts == 'logs': + if os.path.exists(temp_dir): + tarlog = archive_logs(d,temp_dir,False) + else: + return + + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[tarlog]) + + else: + store_package(d,tarlog) + +def dumpdata(d): + '''dump environment to "${P}-${PR}.showdata.dump" including all kinds of variables and functions when running a task''' + workdir = bb.data.getVar('WORKDIR', d, 1) + distro = bb.data.getVar('DISTRO', d, 1) + s = d.getVar('S', True) + pf = d.getVar('PF', True) + target_sys = d.getVar('TARGET_SYS', True) + licenses = get_licenses(d) + dumpdir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + if not os.path.exists(dumpdir): + bb.mkdirhier(dumpdir) + + dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump",d)) + + bb.note("Dumping metadata into '%s'" % dumpfile) + f = open(dumpfile, "w") + # emit variables and shell functions + bb.data.emit_env(f, d, True) + # emit the metadata which isnt valid shell + for e in d.keys(): + if bb.data.getVarFlag(e, 'python', d): + f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1))) + f.close() + +def create_diff_gz(d): + '''creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.g gz for mapping all content in 's' including patches to xxx.diff.gz''' + import shutil + + work_dir = d.getVar('WORKDIR', True) + exclude_from = d.getVar('ARCHIVE_EXCLUDE_FROM', True).split() + pf = d.getVar('PF', True) + licenses = get_licenses(d) + target_sys = d.getVar('TARGET_SYS', True) + diff_dir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + diff_file = os.path.join(diff_dir, bb.data.expand("${P}-${PR}.diff.gz",d)) + + f = open(os.path.join(work_dir,'temp/exclude-from-file'), 'a') + for i in exclude_from: + f.write(i) + f.write("\n") + f.close() + + s=d.getVar('S', True) + distro = d.getVar('DISTRO',True) + dest = s + '/' + distro + '/files' + if not os.path.exists(dest): + bb.mkdirhier(dest) + for i in os.listdir(os.getcwd()): + if os.path.isfile(i): + try: + shutil.copy(i, dest) + except IOError: + os.system('fakeroot cp -rf ' + i + " " + dest ) + + bb.note("Creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.gz") + cmd = "LC_ALL=C TZ=UTC0 diff --exclude-from=" + work_dir + "/temp/exclude-from-file -Naur " + s + '.org' + ' ' + s + " | gzip -c > " + diff_file + d.setVar('DIFF', cmd + "\n") + d.setVarFlag('DIFF', 'func', '1') + bb.build.exec_func('DIFF', d) + shutil.rmtree(s + '.org', ignore_errors=True) + +# This function will run when user want to get tarball for sources and patches after do_unpack +python do_archive_original_sources_patches(){ + archive_sources_patches(d,'prepatch') +} + +# This function will run when user want to get tarball for patched sources after do_patch +python do_archive_patched_sources(){ + archive_sources_patches(d,'patched') +} + +# This function will run when user want to get tarball for configured sources after do_configure +python do_archive_configured_sources(){ + archive_sources_patches(d,'configured') +} + +# This function will run when user want to get tarball for logs or both logs and scripts(.bb and .inc files) +python do_archive_scripts_logs(){ + archive_scripts_logs(d) +} + +# This function will run when user want to know what variable and functions in a running task are and also can get a diff file including +# all content a package should include. +python do_dumpdata_create_diff_gz(){ + dumpdata(d) + create_diff_gz(d) +} + +# This functions prepare for archiving "linux-yocto" because this package create directory 's' before do_patch instead of after do_unpack. +# This is special control for archiving linux-yocto only. +python do_archive_linux_yocto(){ + s = d.getVar('S', True) + if 'linux-yocto' in s: + source_tar_name = archive_sources(d,'') + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[source_tar_name,'']) +} +do_kernel_checkout[postfuncs] += "do_archive_linux_yocto " + +# remove tarball for sources, patches and logs after creating srpm. +python do_remove_tarball(){ + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() == 'SRPM': + work_dir = d.getVar('WORKDIR', True) + try: + for file in os.listdir(os.getcwd()): + if file in get_package(d): + os.remove(file) + os.remove(os.path.join(work_dir,'tar-package')) + except (TypeError,OSError): + pass +} +do_remove_taball[deptask] = "do_archive_scripts_logs" +do_package_write_rpm[postfuncs] += "do_remove_tarball " +export get_licenses +export get_package