A cool command line tool which downloads all images in the given webpage.
Build Status | Version | Downloads |
---|---|---|
###tar file: Grab the latest build using - Pip: https://pypi.python.org/pypi/ImageScraper
###pip install You can also download using pip:
$ pip install ImageScraper
####Dependencies
Note that ImageScraper
depends on lxml
and requests
.
If you run into problems in the compilation of lxml
through pip
, install the libxml2-dev
and libxslt-dev
packages on your system.
$ image-scraper [OPTIONS] URL
-h, --help Print help
-m, --max-images <number> Maximum number images to be scraped
-s, --save-dir <path> Name of the folder to save the images (default: ./images_<domain>)
--max-filesize <size> Limit on size of image in bytes (default: 100000000)
--dump-urls Print the URLs of the images
###If you downloaded the tar: Extract the contents of the tar file.
$cd ImageScraper/
$python setup.py install
$image-scraper --max-images 10 [url to scrape]
###If installed using pip: Open python in terminal.
$image-scraper --max-images 10 [url to scrape]
####NOTE: A new folder called "images_" will be created in the same place, containing all the downloaded images.
Check if a newer version if available and upgrade using:
$ sudo pip install ImageScraper --upgrade
Q.)All images were not downloaded?
It could be that the content was injected into the page via javascript and this scraper doesn't run javascript.
If you want to add features, improve them, or report issues, feel free to send a pull request!!
###Contributors