[Buildroot] [PATCH 1/3] scanpypi.py: new utility
Arnout Vandecappelle
arnout at mind.be
Sun Jun 21 12:29:16 UTC 2015
On 06/15/15 12:06, Denis THULIN wrote:
> Signed-off-by: Denis THULIN <denis.thulin at openwide.fr>
> ---
> v0: initial commit
> python-pacakage-generator.py is an utility for automatically generating a
> python package. It fetches packages info from http://pypi.python.org and
> generates corresponding packages files.
This should go above your Sob so there's an actual commit message.
>
> v1:
> - renamed python-package-generator to scanpypi
> - split the huge script into a lot of functions
> - fixed mistakes and small bugs
>
> I did not know where to put the script so I put it in support/scripts.
> I have updated the python-package section of the manual as well.
We generally make the update to the manual a separate patch, but this is OK for
me as well.
>
> Signed-off-by: Denis THULIN <denis.thulin at openwide.fr>
> ---
> docs/manual/adding-packages-python.txt | 36 ++
> support/scripts/scanpypi.py | 607 +++++++++++++++++++++++++++++++++
> 2 files changed, 643 insertions(+)
> create mode 100755 support/scripts/scanpypi.py
>
> diff --git a/docs/manual/adding-packages-python.txt b/docs/manual/adding-packages-python.txt
> index f81d625..647cb67 100644
> --- a/docs/manual/adding-packages-python.txt
> +++ b/docs/manual/adding-packages-python.txt
> @@ -7,6 +7,42 @@ This infrastructure applies to Python packages that use the standard
> Python setuptools mechanism as their build system, generally
> recognizable by the usage of a +setup.py+ script.
>
> +[[scanpypi]]
> +
> +==== generating a +python-package+ from a pypi repository
Capitalization: Generating
> +
> +You may want to use the +scanpypi.py+ located in
> ++support/script+ to generate a package from an existing pypi(pip) package.
> +
> +you can find the list of existing pypi package here: (https://pypi.python.org).
You
The parenthesis are redundant.
> +
> +Please keep in mind that you most likely need
> +to manually check the package for any mistakes
> +as there are things that cannot be guessed by the generator (e.g.
> +dependencies on any of the python core modules
> +such as BR2_PACKAGE_PYTHON_ZLIB).
Could you rewrap this paragraph? And remove the end-of-line spaces.
> +
> +When at the root of your buildroot directory just do :
> +
> +-----------------------
> +./support/script/scanpypi.py foo bar -o package
> +-----------------------
> +
> +This will generate packages +python-foo+ and +python-bar+ in the package
> +folder if they exist on https://pypi.python.org.
> +
> +You will need to manually write the path to the package inside
> +the +package/Config.in+ file:
: -> .
But perhaps reformulate:
You need to manually add the package to the +package/Config.in+ file.
It would also be better if this sentence was part of the paragraph a couple of
lines higher, were you say that the package has to be checked manually. And the
following sentence should be part of the same paragraph.
> +
> +Find the +external python modules+ menu and insert your package inside.
> +Keep in mind that the items inside a menu should be in alphabetical order.
> +
> +Option +-h+ wil list the options available
> +
> +-----------------------
> +./support/script/scanpypi.py -h
> +-----------------------
> +
> [[python-package-tutorial]]
>
> ==== +python-package+ tutorial
> diff --git a/support/scripts/scanpypi.py b/support/scripts/scanpypi.py
> new file mode 100755
> index 0000000..953f8d2
> --- /dev/null
> +++ b/support/scripts/scanpypi.py
> @@ -0,0 +1,607 @@
> +#!/usr/bin/python2
> +"""
> +Utility for building buildroot packages for existing pypi packages
> +
> +Any package built by scanpypi should be manually checked for
> +errors.
> +"""
> +from __future__ import print_function
> +import argparse
> +import json
> +import urllib2
> +import sys
> +import os
> +import shutil
> +import StringIO
> +import tarfile
> +import errno
> +import hashlib
> +import re
> +import magic
> +import tempfile
> +from functools import wraps
> +
> +
> +# private global
> +_calls = {}
> +
> +
> +def setup_info(pkg_name):
> + """Get a package info from _calls
> +
> + Keyword arguments:
> + pkg_name -- the name of the package
> + """
> + return _calls[pkg_name]
> +
> +
> +def setup_decorator(func, method):
> + """
> + Decorator for distutils.core.setup and setuptools.setup.
> + Puts the args of setup as a dict inside global private dict _calls.
> + Add key 'method' which should be either 'setuptools' or 'distutils'.
> +
> + Keyword arguments:
> + func -- either setuptools.setup or distutils.core.setup
> + method -- either 'setuptools' or 'distutils'
> + """
> +
> + @wraps(func)
> + def closure(*args, **kwargs):
> + _calls[kwargs['name']] = kwargs
> + _calls[kwargs['name']]['method'] = method
> + return closure
> +
> +
> +# monkey patch
> +import setuptools
> +setuptools.setup = setup_decorator(setuptools.setup, 'setuptools')
> +import distutils
> +distutils.core.setup = setup_decorator(setuptools.setup, 'distutils')
> +
> +
> +def find_file_upper_case(filenames, path='./'):
> + """
> + List generator:
> + Recursively find files that matches one of the specified filenames.
> + Returns absolute path
> +
> + Keyword arguments:
> + filenames -- List of filenames to be found
> + path -- Path to the directory to search
> + """
> + for root, dirs, files in os.walk(path):
> + for file in files:
> + if file.upper() in filenames:
> + yield (os.path.join(root, file))
> +
> +
> +def pkg_buildroot_name(pkg_name):
> + """
> + Returns name to avoid troublesome characters.
> + Remove all non alphanumeric characters except -
> + Also lowers the name
> +
> + Keyword arguments:
> + pkg_name -- String to rename
> + """
> + name = re.sub('[^\w-]', '', pkg_name.lower())
> + name = re.sub('^python-', '', name)
> + return name
> +
> +
> +def find_setup(package_name, version, archive):
> + """
> + Search for setup.py file in an archive and returns True if found
> + Used for finding the correct path to the setup.py
> +
> + Keyword arguments:
> + package_name -- base name of the package to search (e.g. Flask)
> + version -- version of the package to search (e.g. 0.8.1)
> + archive -- tar archive to search in
> + """
> + try:
> + archive.getmember('{name}-{version}/setup.py'.format(
> + name=package_name,
> + version=version))
> + except KeyError:
> + return False
> + else:
> + return True
> +
> +
> +def fetch_package_info(pkg_name):
> + """
> + Fetch a package's metadata for the python package index
> +
> + Keyword arguments:
> + pkg_name -- the name of the package
> + """
> + url = 'https://pypi.python.org/pypi/{pkg}/json'.format(
> + pkg=pkg_name)
> + print('URL:', url)
> + try:
> + pkg_json = urllib2.urlopen(url).read().decode()
> + except (urllib2.HTTPError) as error:
I don't think these parenthesis are needed?
> + print('ERROR:', error.getcode(), error.msg, file=sys.stderr)
> + print('ERROR: Could not find package {pkg}.\n'
> + 'Check syntax inside the python package index:\n'
> + 'https://pypi.python.org/pypi/ '.format(pkg=pkg_name))
> + return None, None
> + except urllib2.URLError:
> + print('ERROR: Could not find package {pkg}.\n'
> + 'Check syntax inside the python package index:\n'
> + 'https://pypi.python.org/pypi/ '.format(pkg=pkg_name))
> + return None, None
> +
> + else:
> + return pkg_json, url
> +
> +
> +def download_package(package):
> + """
> + Download a package using metadata from pypi
> +
> + Keyword arguments:
> + package -- a dictionary containing info from the pypi json api
> + """
> + try:
> + targz = package['urls'][0]['filename']
> + except IndexError:
> + print(
> + 'Non conventional package, ',
> + 'please check manually after creation')
Is it even worthwhile to support this case?
> + download_url = package['info']['download_url']
> + try:
> + download = urllib2.urlopen(download_url)
> + except urllib2.HTTPError:
Shouldn't we print an error message here? Or does urrlib already do that?
> + targz = None
> + download = None
> + as_file = None
> + used_url = None
> + else:
> + used_url = {'url': download_url}
> + as_file = StringIO.StringIO(download.read())
Actually, you use the download mostly as a string, it's only used as a file in
tarfile.open(). So I'd put the StringIO call there, and keep it as a string
everywhere else.
Also, perhaps it's worthwhile to make a class for the package information
instead of passing around tuples? I'm not entirely sure if that really makes
things simpler, but it's something to consider.
> + as_file.seek(0)
> + extension = 'tar.gz'
> + if 'gzip' not in magic.from_buffer(as_file.read()):
> + extension = 'tar.bz2'
This part I really don't like. Can't we just get the extension from the URL?
> + targz = '{name}-{version}.{extension}'.format(
> + name=package['info']['name'],
> + version=package['info']['version'], extension=extension)
> + as_file.seek(0)
> + used_url['filename'] = targz
> + else:
> + for download_url in package['urls']:
> + try:
> + download = urllib2.urlopen(download_url['url'])
> + except urllib2.HTTPError:
> + targz = None
> + download = None
> + as_file = None
> + used_url = None
> + else:
> + used_url = download_url
> + as_file = StringIO.StringIO(download.read())
> + md5_sum = hashlib.md5(as_file.read()).hexdigest()
> + if md5_sum == download_url['md5_digest']:
> + break
> + targz = used_url['filename']
> + return download, targz, used_url, as_file
> +
> +
> +def extract_package(pkg_name, as_file, tmp_path):
> + """
> + Create folders used for extracting a package as file object and extract it
> +
> + pkg_name -- name of the package to be extracted
> + as_file -- file object to extract
> + tmp_path -- folder where you want the package to be extracted
> + """
> + as_file.seek(0)
> + as_tarfile = tarfile.open(fileobj=as_file)
Better use a with clause here.
> + tmp_pkg = '/'.join([tmp_path, pkg_name])
os.path.join is better I think.
> + try:
> + os.makedirs(tmp_pkg)
> + except OSError as exception:
> + if exception.errno != errno.EEXIST:
> + print("ERROR: ", exception.message, file=sys.stderr)
> + return None, None
> + print('WARNING:', exception.message, file=sys.stderr)
> + print('Removing {pkg}...'.format(pkg=tmp_pkg))
> + shutil.rmtree(tmp_pkg)
> + os.makedirs(tmp_pkg)
> + version = package['info']['version']
This accidentally works because the package variable is set in the global
scope, but I think it's better to pass this as an argument to the function. Or
better yet, make a class :-)
> + tar_folder = package['info']['name']
> + if not find_setup(tar_folder, version, as_tarfile):
> + return None, None
Since the tarball will anyway be extracted below, perhaps it's easier to do
this find_setup on the extracted directory.
> + as_tarfile.extractall(tmp_pkg)
> + as_tarfile.close()
> + as_file.close()
> + tmp_extract = '{folder}/{name}-{version}'.format(
> + folder=tmp_pkg,
> + name=tar_folder,
> + version=package['info']['version'])
> + return tar_folder, tmp_extract
> +
> +
> +def get_requirements(package_name):
> + """
> + Retrieve dependencies of from a metadata found in the setup.py script of
^^^^^^^^^
dependencies from the metadata
> + a pypi package.
> +
> + Keyword Arguments:
> + package_name -- name of the package found in the pypi metadata of the
> + package.
> + """
> + pkg_req = setup_info(package_name)['install_requires']
> + pkg_req = [re.sub('([\w-]+)[><=]*.*', r'\1', req).lower()
> + for req in pkg_req]
> + pkg_req = map(pkg_buildroot_name, pkg_req)
> + req_not_found = [
> + pkg for pkg in pkg_req
> + if 'python-{name}'.format(name=pkg)
> + not in os.listdir(pkg_folder)
> + ]
> + req_not_found = [pkg for pkg in req_not_found
> + if pkg not in packages]
> + if (req_not_found) != 0:
Just "if req_not_found:" - it's a list, so comparing to 0 is weird.
> + print(
> + 'Error: could not find packages \'{packages}\''
> + 'required by {current_package}'.format(
> + packages=", ".join(req_not_found),
> + current_package=pkg_name))
scancpan instead adds the dependencies to the list of packages to create. We
could do something like this here as well. But that can be done in a follow-up
patch.
> + return pkg_req
> +
> +
> +def create_mk_header(pkg_name):
> + """
> + Create the header of the <package_name>.mk file
> +
> + Keyword arguments:
> + pkg_name -- name of the package
> + """
> + header = ['#' * 80 + '\n']
> + header.append('#\n')
> + header.append('# python-{name}\n'.format(name=pkg_name))
> + header.append('#\n')
> + header.append('#' * 80 + '\n')
> + header.append('\n')
> + return header
I would return a string here instead of a list, so it can all be concatenated
easily in the calling function.
> +
> +
> +def create_mk_download_info(pkg_name, version, targz, url):
> + """
> + Create the lines refering to the download information of the
> + <package_name>.mk file
> +
> + Keyword arguments:
> + pkg_name -- name of the package
> + version -- version of the package
> + targz -- name of the archive corresponding to the package
> + url -- url to be used for downloading the package
> + """
> + lines = []
> + version_line = 'PYTHON_{name}_VERSION = {version}\n'.format(
> + name=pkg_name.upper(),
It's not just .upper(), also - has to replaced with _
> + version=version)
> + lines.append(version_line)
> +
> + targz = targz.replace(
> + version,
> + '$(PYTHON_{name}_VERSION)'.format(name=pkg_name.upper()))
> + targz_line = 'PYTHON_{name}_SOURCE = {filename}\n'.format(
> + name=pkg_name.upper(),
> + filename=targz)
> + lines.append(targz_line)
> +
> + site_line = ('PYTHON_{name}_SITE = {url}\n'.format(
> + name=pkg_name.upper(),
> + url=url['url'].replace(url['filename'], '')))
> + if 'sourceforge' in site_line:
> + site_line = ('PYTHON_{name}_SITE = {url}\n'.format(
> + name=pkg_name.upper(),
> + url=url['url']))
> + lines.append(site_line)
> + return lines
> +
> +
> +def create_mk_setup(pkg_name, tar_folder):
> + """
> + Create the line refering to the setup method of the package of the
> + <package_name>.mk file
> +
> + There are two things you can use to make an installer
> + for a python package: distutils or setuptools
> + distutils comes with python but does not support dependancies.
dependencies
> + distutils is mostly still there for backward support.
> + setuptools is what smart people use,
> + but it is not shipped with python :(
> +
> + Keyword Arguments:
> + pkg_name -- name of the package
> + tar_folder -- name of the folder where the setup.py can be found
> + """
> + lines = []
> + setup_type_line = 'PYTHON_{name}_SETUP_TYPE = {method}\n'.format(
> + name=pkg_name.upper(),
> + method=setup_info(tar_folder)['method'])
> + lines.append(setup_type_line)
> + return lines
> +
> +
> +def create_mk_license(pkg_name, license_name, package_location):
> + """
> + Create the lines referring to the package's license informations of the
> + <package_name>.mk file
> +
> + The license's files are found by searching the package for files named
> + license or license.txt (case insensitive).
> + If more than one license file is found, the user is asked to select which
> + ones he wants to use.
> +
> + Keyword Arguments:
> + pkg_name -- name of the package
> + license_name -- name of the license
> + package_location -- where to look for the licenses
> + """
> + lines = []
> + license_line = 'PYTHON_{name}_LICENSE = {license}\n'.format(
> + name=pkg_name.upper(),
> + license=license_name)
> + lines.append(license_line)
> + print('WARNING: License has been set to "{license}",'
> + ' please change it manually if necessary'.format(
> + license=license_name))
> + filenames = ['LICENSE', 'LICENSE.TXT']
> + license_files = list(find_file_upper_case(filenames, package_location))
> + license_files = [license.replace(package_location, '')[1:]
> + for license in license_files]
> + if len(license_files) > 1:
> + print('More than one file found for license: ')
No need to go interactive here: just make a space-separated list of license
files. You could spew a warning if it is more than one, but it's anyway going to
be verified manually.
> + for index, item in enumerate(license_files):
> + print('\t{index})'.format(index=index), item)
> + license_choices = raw_input(
> + 'specify file numbers separated by spaces(default 0): ')
> + license_choices = [int(choice)
> + for choice in license_choices.split(' ')
> + if choice.isdigit() and int(choice) in
> + range(len(license_files))]
> + if len(license_choices) == 0:
> + license_choices = [0]
> + license_files = [file
> + for index, file in enumerate(license_files)
> + if index in license_choices]
> + elif len(license_files) == 0:
> + print('WARNING: No license file found,'
> + ' please specify it manually afterward')
> +
> + license_file_line = ('PYTHON_{name}_LICENSE_FILES ='
This line should not be added if len(license_files) == 0, but instead there
should be something like:
# No license file found
> + ' {files}\n'.format(
> + name=pkg_name.upper(),
> + files=' '.join(license_files)))
> + license_file_line = license_file_line.replace(' \n', '\n')
If len(license_files) > 0 then this will not be needed.
> + lines.append(license_file_line)
> + return lines
> +
> +
> +def create_mk_requirements(pkg_name, pkg_req):
> + """
> + Create the lines referring to the dependencies of the of the
> + <package_name>.mk file
> +
> + Keyword Arguments:
> + pkg_name -- name of the package
> + pkg_req -- dependencies of the package
> + """
> + lines = []
> + python_pkg_req = ['python-{name}'.format(name=pkg)
> + for pkg in pkg_req]
> + dependencies_line = ('PYTHON_{name}_DEPENDENCIES ='
> + ' {reqs}\n'.format(
> + name=pkg_name.upper(),
> + reqs=' '.join(python_pkg_req)))
> + lines.append(dependencies_line)
> + return lines
> +
> +
> +def create_config_mk(pkg_name, version, license, url, targz,
> + tar_folder, pkg_req, package_location):
create_config_mk -> create_package_mk
> + """
> + Create the lines corresponding to the <package_name>.mk file
> +
> + Keyword Arguments:
> + pkg_name -- name of the package
> + version -- version of the package
> + license -- name of the package's license
> + url -- where to download the package
> + targz -- name of the archive when downloaded
> + tar_folder -- name of the folder where the setup.py can be found
> + pkg_req -- dependencies of the package
> + package_location -- path to the extracted package
> + """
> + lines = create_mk_header(pkg_name)
> + lines += create_mk_download_info(pkg_name, version, targz, url)
> + lines += create_mk_setup(pkg_name, tar_folder)
> + lines += create_mk_license(pkg_name, license, package_location)
> + if pkg_req:
> + lines += create_mk_requirements(pkg_name, pkg_req)
> +
> + lines.append('\n')
> + lines.append('$(eval $(python-package))')
> + lines.append('\n')
> +
> + return lines
> +
> +
> +def create_hash_file(url, digest, hash_function='sha356'):
^^^^^^sha256
> + """
> + Create the lines corresponding to the <package_name>.hash files
> +
> + Keyword Arguments:
> + url -- metadata 'url' from the pypi json api
> + digest -- digest made from the downladed archive
> + hash_function -- algorythm used for hashing
algorithm
> + """
> + lines = []
> + commented_line = '# {method} calculated by scanpypi\n'.format(
I think there would typically be one md5 that comes from pypi, and one sha256
that is calculated locally. So the comment is not correct.
> + method=hash_function)
> + lines.append(commented_line)
> + hash_line = '{method}\t{digest} {filename}\n'.format(
> + method=hash_function,
> + digest=digest,
> + filename=url['filename'])
> + lines.append(hash_line)
> + return lines
> +
> +
> +def create_config_in(pkg_name, pkg_req, package):
> + """
> + Creates the Config.in file of a package
> +
> + pkg_name -- name of the package
> + pkg_req -- dependencies of the package
> + package -- metadata of the package from pypi
> + """
> + lines = []
> + config_line = 'config BR2_PACKAGE_PYTHON_{name}\n'.format(
> + name=pkg_name.upper())
> + lines.append(config_line)
> + python_line = '\tdepends on BR2_PACKAGE_PYTHON\n'
Why? Are all pypi packages python2-only? Remember, the python2||python3
condition is already in package/Config.in.
> + lines.append(python_line)
> +
> + bool_line = '\tbool "python-{name}"\n'.format(name=pkg_name)
> + lines.append(bool_line)
> + if pkg_req:
> + for dep in pkg_req:
> + dep_line = '\tselect BR2_PACKAGE_PYTHON_{req}\n'.format(
> + req=dep.upper())
> + lines.append(dep_line)
> +
> + lines.append('\thelp\n')
> +
> + help_lines = package['info']['summary'].split('\n')
The help_lines should also be wrapped, so call textwrap.wrap() on it.
> + help_lines.append('')
> + help_lines.append(package['info']['home_page'])
> + help_lines = ['\t {line}\n'.format(line=line)
> + for line in help_lines]
> + lines += help_lines
> + return lines
> +
> +
> +if __name__ == "__main__":
> +
> + # Building the parser
> + parser = argparse.ArgumentParser(
> + description="Creates buildroot packages from the metadata of "
> + "an existing pypi(pip) packages and include it "
> + "in menuconfig")
> + parser.add_argument("packages",
> + help="list of packages to be made",
> + nargs='+')
> + parser.add_argument("-o", "--output",
> + help="""
> + Output directory for packages
> + """,
> + default='.')
Since it will be called from the buildroot top dir, this should default to
./packages I think.
> +
> + args = parser.parse_args()
> + packages = list(set(args.packages))
> +
> + # tmp_path is where we'll extract the files later
> + tmp_prefix = 'scanpypi-'
> + pkg_folder = args.output
> + tmp_path = tempfile.mkdtemp(prefix=tmp_prefix)
There should be a big try: block around here that removes the tmpdir
unconditionally. Clearly that makes debugging harder, but it's easy to remove
the rmtree from the script when you want to debug.
> +
> + for real_pkg_name in packages:
> + pkg_name = pkg_buildroot_name(real_pkg_name)
> + print('buildroot package name for {}:'.format(real_pkg_name),
> + pkg_name)
> + # First we download the package
> + # Most of the info we need can only be found inside the package
> + print('Package:', pkg_name)
> + print('Fetching package', real_pkg_name)
> + pkg_json, url = fetch_package_info(real_pkg_name)
> + if not pkg_json:
> + continue
> +
> + pkg_dir = pkg_folder + '/python-' + pkg_name
> + package = json.loads(pkg_json)
> + used_url = ''
> + print('Downloading package {pkg}...'.format(
> + pkg=package['info']['name']))
> + download, targz, used_url, as_file = download_package(package)
> + version = package['info']['version']
> +
> + if not download:
> + print('Error downloading package :', pkg_name)
> + continue
> +
> + sha256_digest = hashlib.sha256(as_file.read()).hexdigest()
> +
> + # extract the tarball
> + tar_folder, tmp_extract = extract_package(pkg_name, as_file, tmp_path)
> +
> + # Loading the package install info from the package
> + sys.path.append(tmp_extract)
> + print(tmp_extract)
This print is not necessary or useful I think.
> + import setup
> + setup = reload(setup)
A comment explaining why the reload is necessary would be useful.
Also, it seems that some packages import other modules from their source
directory. So instead of appending to sys.path, I think it's better to
temporarily chdir to the tmp_extract directory. Try for instance
json-schema-validator
> + sys.path.remove(tmp_extract)
> +
> + pkg_req = None
> + # Package requierement are an argument of the setup function
requirements
> + if 'install_requires' in setup_info(tar_folder):
> + pkg_req = get_requirements(tar_folder)
> + # We could stop here
> + # or ask the user if he still wants to continue
> +
> + # Buildroot python packages require 3 files
> + # The first is the mk file
> + # See:
> + # http://buildroot.uclibc.org/downloads/manual/manual.html
Indentation is wrong.
But I don't think this comment is very useful.
> + print('Checking if package {name} already exists...'.format(
> + name=pkg_dir))
> + try:
> + os.makedirs(pkg_dir)
> + except OSError as exception:
> + if exception.errno != errno.EEXIST:
> + print("ERROR: ", exception.message, file=sys.stderr)
> + continue
> + print('Error: Package {name} already exists'.format(name=pkg_dir))
> + del_pkg = raw_input(
> + 'Do you want to delete existing package ? [y/N]')
> + if del_pkg.lower() == 'y':
> + shutil.rmtree(pkg_dir)
> + os.makedirs(pkg_dir)
> + else:
> + continue
> + pkg_mk = 'python-{name}.mk'.format(name=pkg_name)
> + path_to_mk = '/'.join([pkg_dir, pkg_mk])
> + print('Creating {file}...'.format(file=path_to_mk))
> + config_mk_lines = create_config_mk(pkg_name, version,
> + package['info']['license'],
> + used_url, targz, tar_folder,
> + pkg_req, tmp_extract)
> + with open(path_to_mk, 'w') as mk_file:
> + mk_file.writelines(config_mk_lines)
I think it's more appropriate to do the file writing inside the function.
> +
> + # The second file we make is the hash file
> + # It consists of hashes of the package tarball
> + # http://buildroot.uclibc.org/downloads/manual/manual.html#adding-packages-hash
> + pkg_hash = 'python-{name}.hash'.format(name=pkg_name)
> + path_to_hash = '/'.join([pkg_dir, pkg_hash])
> + print('Creating {filename}...'.format(filename=path_to_hash))
> + hash_lines = create_hash_file(used_url, sha256_digest)
The pypi md5 should be included as well.
Regards,
Arnout
> + with open(path_to_hash, 'w') as hash_file:
> + hash_file.writelines(hash_lines)
> +
> + # The Config.in is the last file we create
> + # It is used by buildroot's menuconfig, gconfig, xconfig or nconfig
> + # it is used to displayspackage info and to select requirements
> + # http://buildroot.uclibc.org/downloads/manual/manual.html#_literal_config_in_literal_file
> + path_to_config = '/'.join([pkg_dir, 'Config.in'])
> + print('Creating {file}...'.format(file=path_to_config))
> + config_in_lines = create_config_in(pkg_name, pkg_req, package)
> + with open(path_to_config, 'w') as config_file:
> + config_file.writelines(config_in_lines)
>
--
Arnout Vandecappelle arnout at mind be
Senior Embedded Software Architect +32-16-286500
Essensium/Mind http://www.mind.be
G.Geenslaan 9, 3001 Leuven, Belgium BE 872 984 063 RPR Leuven
LinkedIn profile: http://www.linkedin.com/in/arnoutvandecappelle
GPG fingerprint: 7CB5 E4CC 6C2E EFD4 6E3D A754 F963 ECAB 2450 2F1F
More information about the buildroot
mailing list