Youtube crawler to measure end-to-end video reception quality
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
apidoc_sphinx_rst
linux_helpers
mibs
pytomo
.gitignore
DESCRIPTION.html
DESCRIPTION.txt
LICENSE
MANIFEST.in
MANIFEST_windows.in
README.html
README.md
README.txt
__init__.py
doctest_runner.py
googlecode_upload.py
lib_plot.py
pytomo.man
pytomo.spec
setup.py
setup_windows.py
start_crawl.py
start_server.py
test_http_server.py

README.md

PYTOMO: A YouTube Crawler

Pytomo is a YouTube crawler designed to figure out network information out of YouTube video download. Video crawl

pytomo [-r max_rounds] [-u max_crawled_url] [-p max_per_url] [-P
max_per_page] [-t time_frame] [-n ping_packets] [-D download_time] [-B
buffering_video_duration] [-M min_playout_buffer_size] [-x] [-L log_level] [-R]
[input_urls]:

Options:
 -h, --help            show this help message and exit
 -r MAX_ROUNDS         Max number of rounds to perform (default 10000)
 -l, --loop            Loop after completing the max nb of rounds
 -u MAX_CRAWLED_URL    Max number of urls to visit (default 5000000)
 -p MAX_PER_URL        Max number of related urls from each page (default 20)
 -P MAX_PER_PAGE       Max number of related videos from each page (default
                       20)
 -t TIME_FRAME         Timeframe for the most popular videos to fetch at
                       start of crawl put 'today', 'week', 'month' or
                       'all_time' (default 'week')
 -n PING_PACKETS       Number of packets to be sent for each ping (default
                       10)
 -D DOWNLOAD_TIME      Download time for the video in seconds (default
                       30.000000)
 -B INITIAL_PLAYBACK_DURATION¶
                       Buffering video duration in seconds (default
                       2000.000000)
 -M MIN_PLAYOUT_BUFFER_SIZE
                       Minimum Playout Buffer Size in seconds (default
                       0.100000)
 -x                    Do NOT store public IP address of the machine in the
                       logs
 -L LOG_LEVEL          The log level setting for the Logging module.Choose
                       from: 'DEBUG', 'INFO', 'WARNING', 'ERROR' and
                       'CRITICAL' (default 'DEBUG')
 --http-proxy=PROXIES  in case of http proxy to reach Internet (default None)
 --provider=PROVIDER   Indicate the ISP
 -R, --no-related      Do NOT crawl related videos (stays with the first urls
                       found: either most popular or arguments given)

Graphical web interface to display graphs about the crawled data

start_server.py [-v] [-f database] [-d database_directory] port

Options:
 -h, --help            show this help message and exit
 -v, --verbose         run as verbose mode
 -f DB_NAME, --file=DB_NAME
                       run on a specific database (by default the latest
                       database in the default database directory is
                       selected)
 -d DB_DIR_NAME, --dir=DB_DIR_NAME
                       run on a specific directory where the latest database
                       will be selected

Installation-free

In order to provide installation-free package, we provide binary executables for Linux (32 and 64bits), Windows, and Mac OS X. The binaries files were generated with .Pyinstaller (version 1.5RC1).

If you have Python installed, you can directly run the start_crawl.py script at root or the pytomo script in bin directory.

If you have Python installed and the RRDtool and webpy python bindings, you can directly run the 'start_server.py' script at root - you need to specify a port on which to start the server and then connect to that port on the respective host from your favourite browser. For example, you start the graphical web interface on port 5555:

   ./start_server.py 5555

You will then connect on the respective machine (you need to know its IP and make sure it allows external connections, otherwise you will only be able to see the graphs locally on that machine):

   http://127.0.0.1:5555/ <= visualisation on local machine
   http://10.193.224.73:5555/ <= visualisation from remote host

By default, the graphs are collected with information from the latest database in the default database directory. If a new live crawl is started, it is recommended to stop and start again the graphical web interface and refresh the page from your browser.

External Resources

We based the lib_youtube_download on .YouTube Download script: we simplified it at most and include only the classes we needed (and only YouTube video retrieval).

The dns module is taken from the .DNS Python Package: we just modified rdata so that .Pyinstaller include all needed modules.

The extraction of metadata out of video files is an adaptation of .Kaa Metadata Python Package: it has been modified in order to be independent of Kaa-base (thus pure Python and portable).

The graphical web interface requires the following external resources, that we recommend you to install from the repository of your OS distribution:

  1. .RRDtool python binding
  • For Debian based OS: sudo apt-get install python-rrdtool
  • For RHEL based OS: yum install python-rrdtool
  1. .webpy
  • For Debian based OS: sudo apt-get install python-webpy
  • For RHEL based OS:yum install python-webpy

External Links