Skip to content

HTTPS clone URL

Subversion checkout URL

You can clone with
or
.
Download ZIP
Tree: 2625956432
Fetching contributors…

Cannot retrieve contributors at this time

65 lines (54 sloc) 2.043 kB
#!/usr/bin/python
#
# gPodder dependency installer for running the CLI from the source tree
#
# Run "python localdepends.py" and it will download and inject dependencies,
# so you only need a standard Python installation for the command-line utility
#
# Thomas Perl <thp.io/about>; 2012-02-11
#
import urllib2
import re
import sys
import StringIO
import tarfile
import os
import shutil
import tempfile
sys.stdout = sys.stderr
src_dir = os.path.abspath(os.path.join(os.path.dirname(__file__), '..', 'src'))
tmp_dir = tempfile.mkdtemp()
MODULES = [
# Module name, Regex-file chooser (1st group = location in "src/")
('feedparser', r'feedparser-[0-9.]+/feedparser/(feedparser.py)'),
('mygpoclient', r'mygpoclient-[0-9.]+/(mygpoclient/[^/]*\.py)')
]
def get_tarball_url(modulename):
url = 'http://pypi.python.org/pypi/' + modulename
html = urllib2.urlopen(url).read()
match = re.search(r'(http://[^>]*%s-([0-9.]*)\.tar\.gz)' % modulename, html)
return match.group(0) if match is not None else None
for module, required_files in MODULES:
print 'Fetching', module, '...',
tarball_url = get_tarball_url(module)
if tarball_url is None:
print 'Cannot determine download URL for', module, '- aborting!'
break
data = urllib2.urlopen(tarball_url).read()
print '%d KiB' % (len(data)/1024)
tar = tarfile.open(fileobj=StringIO.StringIO(data))
for name in tar.getnames():
match = re.match(required_files, name)
if match is not None:
target_name = match.group(1)
target_file = os.path.join(src_dir, target_name)
if os.path.exists(target_file):
print 'Skipping:', target_file
continue
target_dir = os.path.dirname(target_file)
if not os.path.isdir(target_dir):
os.mkdir(target_dir)
print 'Extracting:', target_name
tar.extract(name, tmp_dir)
shutil.move(os.path.join(tmp_dir, name), target_file)
shutil.rmtree(tmp_dir)
Jump to Line
Something went wrong with that request. Please try again.