Skip to content
Switch branches/tags
Go to file
Cannot retrieve contributors at this time

These Linux scripts scrape data from Wikipedia page about S&P500.

They run with travis-ci every day or at each commit to update the data.

They have been tested under Debian Jessy.

Run the scripts

Install the dependencies

The scripts work with some python and shell scripts glued together with a Makefile.

Install the required python libraries :

cd scripts
pip install -r requirements.txt

You can also work on a virtualenv .

Make the package and publish it

The purpose of the project is to compute the datapackage, to test it and to publish it to a git repository :


Only make the package locally and test it

If you work on the code, you might want to skip to publish step :

make valid.txt