diff options
author | Patrick Williams <patrick@stwcx.xyz> | 2016-08-17 22:31:25 +0300 |
---|---|---|
committer | Patrick Williams <patrick@stwcx.xyz> | 2016-08-22 19:43:26 +0300 |
commit | 60f9d69e016b11c468c98ea75ba0a60c44afbbc4 (patch) | |
tree | ecb49581a9e41a37943c22cd9ef3f63451b20ee7 /yocto-poky/meta/lib/oe/distro_check.py | |
parent | e18c61205e0234b03697129c20cc69c9b3940efc (diff) | |
download | openbmc-60f9d69e016b11c468c98ea75ba0a60c44afbbc4.tar.xz |
yocto-poky: Move to import-layers subdir
We are going to import additional layers, so create a subdir to
hold all of the layers that we import with git-subtree.
Change-Id: I6f732153a22be8ca663035c518837e3cc5ec0799
Signed-off-by: Patrick Williams <patrick@stwcx.xyz>
Diffstat (limited to 'yocto-poky/meta/lib/oe/distro_check.py')
-rw-r--r-- | yocto-poky/meta/lib/oe/distro_check.py | 399 |
1 files changed, 0 insertions, 399 deletions
diff --git a/yocto-poky/meta/lib/oe/distro_check.py b/yocto-poky/meta/lib/oe/distro_check.py deleted file mode 100644 index 8655a6fc1..000000000 --- a/yocto-poky/meta/lib/oe/distro_check.py +++ /dev/null @@ -1,399 +0,0 @@ -from contextlib import contextmanager -@contextmanager -def create_socket(url, d): - import urllib - socket = urllib.urlopen(url, proxies=get_proxies(d)) - try: - yield socket - finally: - socket.close() - -def get_proxies(d): - proxies = {} - for key in ['http', 'https', 'ftp', 'ftps', 'no', 'all']: - proxy = d.getVar(key + '_proxy', True) - if proxy: - proxies[key] = proxy - return proxies - -def get_links_from_url(url, d): - "Return all the href links found on the web location" - - import sgmllib - - class LinksParser(sgmllib.SGMLParser): - def parse(self, s): - "Parse the given string 's'." - self.feed(s) - self.close() - - def __init__(self, verbose=0): - "Initialise an object passing 'verbose' to the superclass." - sgmllib.SGMLParser.__init__(self, verbose) - self.hyperlinks = [] - - def start_a(self, attributes): - "Process a hyperlink and its 'attributes'." - for name, value in attributes: - if name == "href": - self.hyperlinks.append(value.strip('/')) - - def get_hyperlinks(self): - "Return the list of hyperlinks." - return self.hyperlinks - - with create_socket(url,d) as sock: - webpage = sock.read() - - linksparser = LinksParser() - linksparser.parse(webpage) - return linksparser.get_hyperlinks() - -def find_latest_numeric_release(url, d): - "Find the latest listed numeric release on the given url" - max=0 - maxstr="" - for link in get_links_from_url(url, d): - try: - release = float(link) - except: - release = 0 - if release > max: - max = release - maxstr = link - return maxstr - -def is_src_rpm(name): - "Check if the link is pointing to a src.rpm file" - if name[-8:] == ".src.rpm": - return True - else: - return False - -def package_name_from_srpm(srpm): - "Strip out the package name from the src.rpm filename" - strings = srpm.split('-') - package_name = strings[0] - for i in range(1, len (strings) - 1): - str = strings[i] - if not str[0].isdigit(): - package_name += '-' + str - return package_name - -def clean_package_list(package_list): - "Removes multiple entries of packages and sorts the list" - set = {} - map(set.__setitem__, package_list, []) - return set.keys() - - -def get_latest_released_meego_source_package_list(d): - "Returns list of all the name os packages in the latest meego distro" - - package_names = [] - try: - f = open("/tmp/Meego-1.1", "r") - for line in f: - package_names.append(line[:-1] + ":" + "main") # Also strip the '\n' at the end - except IOError: pass - package_list=clean_package_list(package_names) - return "1.0", package_list - -def get_source_package_list_from_url(url, section, d): - "Return a sectioned list of package names from a URL list" - - bb.note("Reading %s: %s" % (url, section)) - links = get_links_from_url(url, d) - srpms = filter(is_src_rpm, links) - names_list = map(package_name_from_srpm, srpms) - - new_pkgs = [] - for pkgs in names_list: - new_pkgs.append(pkgs + ":" + section) - - return new_pkgs - -def get_latest_released_fedora_source_package_list(d): - "Returns list of all the name os packages in the latest fedora distro" - latest = find_latest_numeric_release("http://archive.fedoraproject.org/pub/fedora/linux/releases/", d) - - package_names = get_source_package_list_from_url("http://archive.fedoraproject.org/pub/fedora/linux/releases/%s/Fedora/source/SRPMS/" % latest, "main", d) - -# package_names += get_source_package_list_from_url("http://download.fedora.redhat.com/pub/fedora/linux/releases/%s/Everything/source/SPRMS/" % latest, "everything") - package_names += get_source_package_list_from_url("http://archive.fedoraproject.org/pub/fedora/linux/updates/%s/SRPMS/" % latest, "updates", d) - - package_list=clean_package_list(package_names) - - return latest, package_list - -def get_latest_released_opensuse_source_package_list(d): - "Returns list of all the name os packages in the latest opensuse distro" - latest = find_latest_numeric_release("http://download.opensuse.org/source/distribution/",d) - - package_names = get_source_package_list_from_url("http://download.opensuse.org/source/distribution/%s/repo/oss/suse/src/" % latest, "main", d) - package_names += get_source_package_list_from_url("http://download.opensuse.org/update/%s/rpm/src/" % latest, "updates", d) - - package_list=clean_package_list(package_names) - return latest, package_list - -def get_latest_released_mandriva_source_package_list(d): - "Returns list of all the name os packages in the latest mandriva distro" - latest = find_latest_numeric_release("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/", d) - package_names = get_source_package_list_from_url("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/%s/SRPMS/main/release/" % latest, "main", d) -# package_names += get_source_package_list_from_url("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/%s/SRPMS/contrib/release/" % latest, "contrib") - package_names += get_source_package_list_from_url("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/%s/SRPMS/main/updates/" % latest, "updates", d) - - package_list=clean_package_list(package_names) - return latest, package_list - -def find_latest_debian_release(url, d): - "Find the latest listed debian release on the given url" - - releases = [] - for link in get_links_from_url(url, d): - if link[:6] == "Debian": - if ';' not in link: - releases.append(link) - releases.sort() - try: - return releases.pop()[6:] - except: - return "_NotFound_" - -def get_debian_style_source_package_list(url, section, d): - "Return the list of package-names stored in the debian style Sources.gz file" - with create_socket(url,d) as sock: - webpage = sock.read() - import tempfile - tmpfile = tempfile.NamedTemporaryFile(mode='wb', prefix='oecore.', suffix='.tmp', delete=False) - tmpfilename=tmpfile.name - tmpfile.write(sock.read()) - tmpfile.close() - import gzip - bb.note("Reading %s: %s" % (url, section)) - - f = gzip.open(tmpfilename) - package_names = [] - for line in f: - if line[:9] == "Package: ": - package_names.append(line[9:-1] + ":" + section) # Also strip the '\n' at the end - os.unlink(tmpfilename) - - return package_names - -def get_latest_released_debian_source_package_list(d): - "Returns list of all the name os packages in the latest debian distro" - latest = find_latest_debian_release("http://ftp.debian.org/debian/dists/", d) - url = "http://ftp.debian.org/debian/dists/stable/main/source/Sources.gz" - package_names = get_debian_style_source_package_list(url, "main", d) -# url = "http://ftp.debian.org/debian/dists/stable/contrib/source/Sources.gz" -# package_names += get_debian_style_source_package_list(url, "contrib") - url = "http://ftp.debian.org/debian/dists/stable-proposed-updates/main/source/Sources.gz" - package_names += get_debian_style_source_package_list(url, "updates", d) - package_list=clean_package_list(package_names) - return latest, package_list - -def find_latest_ubuntu_release(url, d): - "Find the latest listed ubuntu release on the given url" - url += "?C=M;O=D" # Descending Sort by Last Modified - for link in get_links_from_url(url, d): - if link[-8:] == "-updates": - return link[:-8] - return "_NotFound_" - -def get_latest_released_ubuntu_source_package_list(d): - "Returns list of all the name os packages in the latest ubuntu distro" - latest = find_latest_ubuntu_release("http://archive.ubuntu.com/ubuntu/dists/", d) - url = "http://archive.ubuntu.com/ubuntu/dists/%s/main/source/Sources.gz" % latest - package_names = get_debian_style_source_package_list(url, "main", d) -# url = "http://archive.ubuntu.com/ubuntu/dists/%s/multiverse/source/Sources.gz" % latest -# package_names += get_debian_style_source_package_list(url, "multiverse") -# url = "http://archive.ubuntu.com/ubuntu/dists/%s/universe/source/Sources.gz" % latest -# package_names += get_debian_style_source_package_list(url, "universe") - url = "http://archive.ubuntu.com/ubuntu/dists/%s-updates/main/source/Sources.gz" % latest - package_names += get_debian_style_source_package_list(url, "updates", d) - package_list=clean_package_list(package_names) - return latest, package_list - -def create_distro_packages_list(distro_check_dir, d): - pkglst_dir = os.path.join(distro_check_dir, "package_lists") - if not os.path.isdir (pkglst_dir): - os.makedirs(pkglst_dir) - # first clear old stuff - for file in os.listdir(pkglst_dir): - os.unlink(os.path.join(pkglst_dir, file)) - - per_distro_functions = [ - ["Debian", get_latest_released_debian_source_package_list], - ["Ubuntu", get_latest_released_ubuntu_source_package_list], - ["Fedora", get_latest_released_fedora_source_package_list], - ["OpenSuSE", get_latest_released_opensuse_source_package_list], - ["Mandriva", get_latest_released_mandriva_source_package_list], - ["Meego", get_latest_released_meego_source_package_list] - ] - - from datetime import datetime - begin = datetime.now() - for distro in per_distro_functions: - name = distro[0] - release, package_list = distro[1](d) - bb.note("Distro: %s, Latest Release: %s, # src packages: %d" % (name, release, len(package_list))) - package_list_file = os.path.join(pkglst_dir, name + "-" + release) - f = open(package_list_file, "w+b") - for pkg in package_list: - f.write(pkg + "\n") - f.close() - end = datetime.now() - delta = end - begin - bb.note("package_list generatiosn took this much time: %d seconds" % delta.seconds) - -def update_distro_data(distro_check_dir, datetime, d): - """ - If distro packages list data is old then rebuild it. - The operations has to be protected by a lock so that - only one thread performes it at a time. - """ - if not os.path.isdir (distro_check_dir): - try: - bb.note ("Making new directory: %s" % distro_check_dir) - os.makedirs (distro_check_dir) - except OSError: - raise Exception('Unable to create directory %s' % (distro_check_dir)) - - - datetime_file = os.path.join(distro_check_dir, "build_datetime") - saved_datetime = "_invalid_" - import fcntl - try: - if not os.path.exists(datetime_file): - open(datetime_file, 'w+b').close() # touch the file so that the next open won't fail - - f = open(datetime_file, "r+b") - fcntl.lockf(f, fcntl.LOCK_EX) - saved_datetime = f.read() - if saved_datetime[0:8] != datetime[0:8]: - bb.note("The build datetime did not match: saved:%s current:%s" % (saved_datetime, datetime)) - bb.note("Regenerating distro package lists") - create_distro_packages_list(distro_check_dir, d) - f.seek(0) - f.write(datetime) - - except OSError: - raise Exception('Unable to read/write this file: %s' % (datetime_file)) - finally: - fcntl.lockf(f, fcntl.LOCK_UN) - f.close() - -def compare_in_distro_packages_list(distro_check_dir, d): - if not os.path.isdir(distro_check_dir): - raise Exception("compare_in_distro_packages_list: invalid distro_check_dir passed") - - localdata = bb.data.createCopy(d) - pkglst_dir = os.path.join(distro_check_dir, "package_lists") - matching_distros = [] - pn = d.getVar('PN', True) - recipe_name = d.getVar('PN', True) - bb.note("Checking: %s" % pn) - - trim_dict = dict({"-native":"-native", "-cross":"-cross", "-initial":"-initial"}) - - if pn.find("-native") != -1: - pnstripped = pn.split("-native") - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - recipe_name = pnstripped[0] - - if pn.startswith("nativesdk-"): - pnstripped = pn.split("nativesdk-") - localdata.setVar('OVERRIDES', "pn-" + pnstripped[1] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - recipe_name = pnstripped[1] - - if pn.find("-cross") != -1: - pnstripped = pn.split("-cross") - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - recipe_name = pnstripped[0] - - if pn.find("-initial") != -1: - pnstripped = pn.split("-initial") - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - recipe_name = pnstripped[0] - - bb.note("Recipe: %s" % recipe_name) - tmp = localdata.getVar('DISTRO_PN_ALIAS', True) - - distro_exceptions = dict({"OE-Core":'OE-Core', "OpenedHand":'OpenedHand', "Intel":'Intel', "Upstream":'Upstream', "Windriver":'Windriver', "OSPDT":'OSPDT Approved', "Poky":'poky'}) - - if tmp: - list = tmp.split(' ') - for str in list: - if str and str.find("=") == -1 and distro_exceptions[str]: - matching_distros.append(str) - - distro_pn_aliases = {} - if tmp: - list = tmp.split(' ') - for str in list: - if str.find("=") != -1: - (dist, pn_alias) = str.split('=') - distro_pn_aliases[dist.strip().lower()] = pn_alias.strip() - - for file in os.listdir(pkglst_dir): - (distro, distro_release) = file.split("-") - f = open(os.path.join(pkglst_dir, file), "rb") - for line in f: - (pkg, section) = line.split(":") - if distro.lower() in distro_pn_aliases: - pn = distro_pn_aliases[distro.lower()] - else: - pn = recipe_name - if pn == pkg: - matching_distros.append(distro + "-" + section[:-1]) # strip the \n at the end - f.close() - break - f.close() - - - if tmp != None: - list = tmp.split(' ') - for item in list: - matching_distros.append(item) - bb.note("Matching: %s" % matching_distros) - return matching_distros - -def create_log_file(d, logname): - import subprocess - logpath = d.getVar('LOG_DIR', True) - bb.utils.mkdirhier(logpath) - logfn, logsuffix = os.path.splitext(logname) - logfile = os.path.join(logpath, "%s.%s%s" % (logfn, d.getVar('DATETIME', True), logsuffix)) - if not os.path.exists(logfile): - slogfile = os.path.join(logpath, logname) - if os.path.exists(slogfile): - os.remove(slogfile) - subprocess.call("touch %s" % logfile, shell=True) - os.symlink(logfile, slogfile) - d.setVar('LOG_FILE', logfile) - return logfile - - -def save_distro_check_result(result, datetime, result_file, d): - pn = d.getVar('PN', True) - logdir = d.getVar('LOG_DIR', True) - if not logdir: - bb.error("LOG_DIR variable is not defined, can't write the distro_check results") - return - if not os.path.isdir(logdir): - os.makedirs(logdir) - line = pn - for i in result: - line = line + "," + i - f = open(result_file, "a") - import fcntl - fcntl.lockf(f, fcntl.LOCK_EX) - f.seek(0, os.SEEK_END) # seek to the end of file - f.write(line + "\n") - fcntl.lockf(f, fcntl.LOCK_UN) - f.close() |