# encoding: utf8 import datetime import os import re import sys import textwrap import pandas as pd import license download_keys = """ Origin Original-Source Source Source-Code X-Origin X-Original-Package X-Source """ def concat(xss): all = [] for xs in xss: all.extend(xs) return all def indent(s): return ''.join( ' %s\n' % line if line else '\n' for line in s.splitlines()) def today(): return datetime.datetime.now().strftime('%Y-%m-%d') def warn(*x): for s in ('warning:',) + x + ('\n',): print >>sys.stderr, s, class ExportFailure(Exception): pass class PkgData(object): def __init__(self): pkg_store = pd.HDFStore('pkg.h5') self.pkgs = pkg_store['packages'] self.descs = pkg_store['descriptions'] pkg_store.close() cp_store = pd.HDFStore('cp.h5') self.cpf = cp_store['cp_files'] self.cps = cp_store['cp_summary'] self.licenses = cp_store['licenses'] cp_store.close() class Template(object): def __init__(self, name, values): self.name = name self.values = values def __str__(self): return '{{%s\n%s\n}}' % ( self.name, '\n'.join(['|' + '%s=%s' % (n, v.encode('utf8') if isinstance(v, unicode) else v) for (n, v) in self.values])) def parse_tags(s): return s.replace('\n', '').split(', ') def extract_languages(tags): langs = [] for tag in tags: (a, b) = tag.split('::') if a == 'implemented-in': langs.append(b) elif a == 'devel' and b.startswith('lang:'): langs.append(b.split(':')[1]) return list(set(langs)) def catechise(s): heresies = ["open source", "debian", "(?]+)>', s) if match: return (match.group(1), match.group(2)) else: return (s, '') def extract_people(df): # XXX: extract contributors, maintainers df = df.dropna() if 'Upstream-Contact' in df: (name, email) = parse_person(df['Upstream-Contact']) yield Template('Person', [ ('Real name', name), ('Role', 'contact'), ('Email', email), ('Resource URL', '')]) def extract_resources(cp_header): cp_header = cp_header.dropna() for key in re.findall('\S+', download_keys): if key in cp_header: yield Template('Resource', [ ('Resource kind', 'Download'), ('Resource URL', cp_header[key])]) def export_srcpkgs(data, name, srcpkg_names): binpkgs = pd.concat([ data.pkgs[data.pkgs['_srcpkg'] == srcpkg] for srcpkg in srcpkg_names]) if len(binpkgs) == 0: warn('no binary packages found for', srcpkg_names) return binpkg_names = sorted(binpkgs['Package'], key=len) homepages = list(binpkgs['Homepage']) # XXX: maybe choose the one that appears the most? homepage = homepages[0] if homepages else '' tags = set(concat( [parse_tags(t) for t in binpkgs['Tag'] if not pd.isnull(t)])) langs = [s.title() for s in extract_languages(tags)] if name in binpkg_names: descpkg = name else: # Heuristic: choose the package with the shortest name. # We could try to do something smarter, like look for the common # prefix of the descriptions of all the binary packages. descpkg = binpkg_names[0] desc = list(data.descs[ data.descs['Package'] == descpkg]['Description-en'])[0] (short_desc, full_desc) = desc.split('\n', 1) full_desc = catechise(munge_description(full_desc)) yield Template('Entry', [ ('Name', name.capitalize()), ('Short description', short_desc), ('Full description', full_desc), ('Homepage URL', homepage), ('User level', ''), # XXX get this information from apt-file ('Component programs', ''), ('VCS checkout command', ''), ('Computer languages', ', '.join(langs)), ('Status', ''), ('Is GNU', 'No'), ('Submitted by', 'Debian import'), ('Submitted date', today())]) yield Template('Import', [ ('Source', 'Debian'), ('Source link', 'http://packages.debian.org/sid/' + srcpkg_names[0]), ('Date', today())]) people = [] res = [] for srcpkg in srcpkg_names: pkg_cps = data.cps[data.cps['_srcpkg'] == srcpkg].ix[0] pkg_cpf = data.cpf[data.cpf['_srcpkg'] == srcpkg] pkg_licenses = data.licenses[data.licenses['_srcpkg'] == srcpkg] people.extend(list(extract_people(pkg_cps))) res.extend(list(extract_resources(pkg_cps))) for template in srcpkg_extract_licenses( pkg_cps, pkg_cpf, pkg_licenses): # XXX: eliminate duplicates yield template for template in people: # XXX: eliminate duplicates yield template for template in res: # XXX: eliminate duplicates yield template #yield Template('Software category', [ # ('Resource kind', ''), # ('Resource URL', '')]) def export(data, name): pkg_cps = data.cps[data.cps['Upstream-Name'] == name] srcpkg_names = list(pkg_cps['_srcpkg']) for template in export_srcpkgs(data, name, srcpkg_names): yield template def filename(s): s_ = re.sub('[^A-Za-z0-9_+.-]', '_', s) assert s_, s return s_ + '.wiki' def output(path, xs): with open(path, 'w') as f: for x in xs: f.write(str(x) + '\n') def try_export(path, name, xs): try: output(path, xs) except ExportFailure, e: warn('export failed: %s: %s' % (name, e.message)) def export_all(data): outputdir = 'output' if not os.path.exists(outputdir): os.makedirs(outputdir) # First, find all upstream names and the source packages corresponding # to them. unames = set(data.cps['Upstream-Name'].dropna()) for uname in unames: if not uname: continue print uname.encode('utf8') fname = os.path.join(outputdir, filename(uname)) try_export(fname, uname, export(data, uname)) # For source packages with no upstream name, use the source package # name as the upstream name. no_uname = set(data.cps[ data.cps['Upstream-Name'].isnull()]['_srcpkg']) for srcpkg in no_uname: print srcpkg fname = os.path.join(outputdir, filename(srcpkg)) try_export(fname, srcpkg, export_srcpkgs(data, srcpkg, [srcpkg])) def main(): data = PkgData() args = sys.argv[1:] if len(args) == 0: export_all(data) elif len(args) == 1: # XXX: assumes argument is an upstream name for template in export(data, args[0]): print template else: raise RuntimeError() if __name__ == '__main__': main()