Fetching contributors…
Cannot retrieve contributors at this time
executable file 86 lines (66 sloc) 2.67 KB
#!/usr/bin/env python
Unfortunately the "expanded_url" as supplied by Twitter aren't fully
expanded one hop past will attempt to completely unshorten URLs and add them as the
"unshortened_url" key to each url, and emit the tweet as JSON again on stdout.
This script starts 10 seaprate processes which talk to an instance of unshrtn
that is running.
from __future__ import print_function
import json
import urllib
import logging
import fileinput
import multiprocessing
import argparse
import time
# number of urls to look up in parallel
unshrtn_url = "http://localhost:3000"
retries = 2
wait = 15
logging.basicConfig(filename="unshorten.log", level=logging.INFO)
def rewrite_line(line):
tweet = json.loads(line)
# don't do the same work again
if 'unshortened_url' in tweet and tweet['unshortened_url']:
return line
for url_dict in tweet["entities"]["urls"]:
if "expanded_url" in url_dict:
url = url_dict["expanded_url"]
url = url_dict['url']
url = url.encode('utf8')
u = '{}/?{}'.format(unshrtn_url, urllib.urlencode({'url': url}))
resp = None
for retry in range(1, retries+1):
resp = json.loads(urllib.urlopen(u).read())
except Exception as e:
logging.error("http error: %s when looking up %s. Try %s of %s", e, url, retry, retries)
if resp and resp['long']:
url_dict['unshortened_url'] = resp['long']
return json.dumps(tweet)
def main():
global unshrtn_url, retries, wait
parser = argparse.ArgumentParser()
parser.add_argument('--pool-size', help='number of urls to look up in parallel', default=POOL_SIZE, type=int)
parser.add_argument('--unshrtn', help='url of the unshrtn service', default=unshrtn_url)
parser.add_argument('--retries', help='number of time to retry if error from unshrtn service', default=retries,
parser.add_argument('--wait', help='number of seconds to wait between retries if error from unshrtn service',
default=wait, type=int)
parser.add_argument('files', metavar='FILE', nargs='*', help='files to read, if empty, stdin is used')
args = parser.parse_args()
unshrtn_url = args.unshrtn
retries = args.retries
wait = args.wait
pool = multiprocessing.Pool(args.pool_size)
for line in pool.imap_unordered(rewrite_line, fileinput.input(files=args.files if len(args.files) > 0 else ('-',))):
if line != "\n": print(line)
if __name__ == "__main__":