summaryrefslogtreecommitdiff
path: root/baserockimport/exts/python.to_lorry
diff options
context:
space:
mode:
authorSam Thursfield <sam.thursfield@codethink.co.uk>2014-12-18 10:43:37 +0000
committerSam Thursfield <sam.thursfield@codethink.co.uk>2014-12-18 10:43:37 +0000
commit764531201f99bf1d9c6dd451a212b741bfb6715e (patch)
treeb9a839ff8cb8000792382805d9027e4891835763 /baserockimport/exts/python.to_lorry
parenteb1a6a511c85163fe3e7ede56a348206075d9af9 (diff)
parent65278fdf1ec80a784f9ada8d390ad063a459c97a (diff)
downloadimport-764531201f99bf1d9c6dd451a212b741bfb6715e.tar.gz
Merge branch 'baserock/richardipsum/python_v3'
There is work still to be done on this importer, but it is usable for some Python projects and may as well be merged to 'master' now.
Diffstat (limited to 'baserockimport/exts/python.to_lorry')
-rwxr-xr-xbaserockimport/exts/python.to_lorry219
1 files changed, 219 insertions, 0 deletions
diff --git a/baserockimport/exts/python.to_lorry b/baserockimport/exts/python.to_lorry
new file mode 100755
index 0000000..b7341ca
--- /dev/null
+++ b/baserockimport/exts/python.to_lorry
@@ -0,0 +1,219 @@
+#!/usr/bin/env python
+#
+# Create a Baserock .lorry file for a given Python package
+#
+# Copyright (C) 2014 Codethink Limited
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; version 2 of the License.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+
+from __future__ import print_function
+
+import subprocess
+import requests
+import json
+import sys
+import shutil
+import tempfile
+import xmlrpclib
+import logging
+import select
+
+import pkg_resources
+
+from importer_python_common import *
+
+def fetch_package_metadata(package_name):
+ try:
+ result = requests.get('%s/%s/json' % (PYPI_URL, package_name))
+
+ # raise exception if status code is not 200 OK
+ result.raise_for_status()
+ except Exception as e:
+ error("Couldn't fetch package metadata:", e)
+
+ return result.json()
+
+def find_repo_type(url):
+
+ # Don't bother with detection if we can't get a 200 OK
+ logging.debug("Getting '%s' ..." % url)
+
+ status_code = requests.get(url).status_code
+ if status_code != 200:
+ logging.debug('Got %d status code from %s, aborting repo detection'
+ % (status_code, url))
+ return None
+
+ logging.debug('200 OK for %s' % url)
+ logging.debug('Finding repo type for %s' % url)
+
+ vcss = [('git', 'clone'), ('hg', 'clone'),
+ ('svn', 'checkout'), ('bzr', 'branch')]
+
+ for (vcs, vcs_command) in vcss:
+ logging.debug('Trying %s %s' % (vcs, vcs_command))
+ tempdir = tempfile.mkdtemp()
+
+ p = subprocess.Popen([vcs, vcs_command, url], stdout=subprocess.PIPE,
+ stderr=subprocess.STDOUT, stdin=subprocess.PIPE,
+ cwd=tempdir)
+
+ # We close stdin on parent side to prevent the child from blocking
+ # if it reads on stdin
+ p.stdin.close()
+
+ while True:
+ line = p.stdout.readline()
+ if line == '':
+ break
+
+ logging.debug(line.rstrip('\n'))
+
+ p.wait() # even with eof on both streams, we still wait
+
+ shutil.rmtree(tempdir)
+
+ if p.returncode == 0:
+ logging.debug('%s is a %s repo' % (url, vcs))
+ return vcs
+
+ logging.debug("%s doesn't seem to be a repo" % url)
+
+ return None
+
+def get_compression(url):
+ bzip = 'bzip2'
+ gzip = 'gzip'
+ lzma = 'lzma'
+
+ m = {'tar.gz': gzip, 'tgz': gzip, 'tar.Z': gzip,
+ 'tar.bz2': bzip, 'tbz2': bzip,
+ 'tar.lzma': lzma, 'tar.xz': lzma, 'tlz': lzma, 'txz': lzma}
+
+ for x in [1, 2]:
+ ext = '.'.join(url.split('.')[-x:])
+ if ext in m: return m[ext]
+
+ return None
+
+# Assumption: url passed to this function must have a 'standard' tar extension
+def make_tarball_lorry(package_name, url):
+ # TODO: this prefix probably shouldn't be hardcoded here either
+ name = 'python-packages/%s' % package_name.lower()
+
+ lorry = {'type': 'tarball', 'url': url}
+ compression = get_compression(url)
+ if compression:
+ lorry['compression'] = compression
+
+ return json.dumps({name + "-tarball": lorry}, indent=4, sort_keys=True)
+
+def filter_urls(urls):
+ allowed_extensions = ['tar.gz', 'tgz', 'tar.Z', 'tar.bz2', 'tbz2',
+ 'tar.lzma', 'tar.xz', 'tlz', 'txz', 'tar']
+
+ def allowed_extension(url):
+ return ('.'.join(url['url'].split('.')[-2:]) in allowed_extensions
+ or url['url'].split('.')[-1:] in allowed_extensions)
+
+ return filter(allowed_extension, urls)
+
+def get_releases(client, requirement):
+ try:
+ releases = client.package_releases(requirement.project_name)
+ except Exception as e:
+ error("Couldn't fetch release data:", e)
+
+ return releases
+
+def generate_tarball_lorry(client, requirement):
+ releases = get_releases(client, requirement)
+
+ if len(releases) == 0:
+ error("Couldn't find any releases for package %s"
+ % requirement.project_name)
+
+ releases = [v for v in releases if specs_satisfied(v, requirement.specs)]
+
+ if len(releases) == 0:
+ error("Couldn't find any releases of %s"
+ " that satisfy version constraints: %s"
+ % (requirement.project_name, requirement.specs))
+
+ release_version = releases[0]
+
+ logging.debug('Fetching urls for package %s with version %s'
+ % (requirement.project_name, release_version))
+
+ try:
+ # Get a list of dicts, the dicts contain the urls.
+ urls = client.release_urls(requirement.project_name, release_version)
+ except Exception as e:
+ error("Couldn't fetch release urls:", e)
+
+ tarball_urls = filter_urls(urls)
+
+ if len(tarball_urls) > 0:
+ urls = tarball_urls
+ elif len(urls) > 0:
+ warn("None of these urls look like tarballs:")
+ for url in urls:
+ warn("\t%s" % url['url'])
+ error("Cannot proceed")
+ else:
+ error("Couldn't find any download urls for package %s"
+ % requirement.project_name)
+
+ url = urls[0]['url']
+
+ return make_tarball_lorry(requirement.project_name, url)
+
+def str_repo_lorry(package_name, repo_type, url):
+ # TODO: this prefix probably shouldn't be hardcoded here
+ name = 'python-packages/%s' % package_name.lower()
+
+ return json.dumps({name: {'type': repo_type, 'url': url}},
+ indent=4, sort_keys=True)
+
+def main():
+ if len(sys.argv) != 2:
+ # TODO explain the format of python requirements
+ # warn the user that they probably want to quote their arg
+ # > < will be interpreted as redirection by the shell
+ print('usage: %s requirement' % sys.argv[0], file=sys.stderr)
+ sys.exit(1)
+
+ client = xmlrpclib.ServerProxy(PYPI_URL)
+
+ req = pkg_resources.parse_requirements(sys.argv[1]).next()
+
+ new_proj_name = name_or_closest(client, req.project_name)
+
+ if new_proj_name == None:
+ error("Couldn't find any project with name '%s'" % req.project_name)
+
+ logging.debug('Treating %s as %s' % (req.project_name, new_proj_name))
+ req.project_name = new_proj_name
+
+ metadata = fetch_package_metadata(req.project_name)
+ info = metadata['info']
+
+ repo_type = (find_repo_type(info['home_page'])
+ if 'home_page' in info else None)
+
+ print(str_repo_lorry(req.project_name, repo_type, info['home_page'])
+ if repo_type else generate_tarball_lorry(client, req))
+
+if __name__ == '__main__':
+ PythonExtension().run()