This is a BASH script to perform fast image downloads sourced from Google Images based on a specified search-phrase. It's a web-page scraper that feeds a list of original image URLs to Wget (or cURL) to download images in parallel, then combine them using ImageMagick's montage into a single gallery image. The idea is to build a picture of a phrase.
This is an expansion upon a solution provided by ShellFish and has been updated to handle Google's various page-code changes from April 2016 to June 2018.
$ wget -qN git.io/googliser.sh && chmod +x googliser.sh
$ curl -skLO git.io/googliser.sh && chmod +x googliser.sh
The user supplies a search-phrase and other optional parameters on the command-line.
A sub-directory with the name of this search-phrase is created below the current directory.
Google Images is queried and the results saved.
The results are parsed and all image links are extracted and saved to a URL list file. Any links for YouTube and Vimeo are removed.
The script iterates through this URL list and downloads the first [n]umber of available images. Up to 1,000 images can be requested. Up to 40 images can be downloaded in parallel (concurrently). If an image is unavailable, it's skipped and downloading continues until the required number of images have been downloaded or the download failure-limit is reached.
googliser is fully supported on Fedora Workstation, Manjaro & Ubuntu. Debian and macOS may require some extra binaries. Please advise of any issues.
$ sudo apt install wget imagemagick
$ xcode-select --install $ ruby -e "$(curl -fsSL git.io/get-brew)" $ brew install coreutils ghostscript gnu-sed gnu-getopt imagemagick
These sample images have been scaled down for easier distribution.
$ ./googliser.sh --phrase "puppies" --title 'Puppies!' --upper-size 100000 --lower-size 2000 --failures 0
$ ./googliser.sh -p "kittens" -T 'Kittens!' -f0 -SC -n16
$ ./googliser.sh -n 380 -p "cows" -u 250000 -l 10000 -f 0 -S
$ ./googliser.sh [PARAMETERS] ...
Allowable parameters are indicated with a hyphen then a single character or the long form with 2 hypens and full-text. Single character options can be concatenated. e.g.
-CdDhLNqsSz. Parameters can be specified as follows:
The search-phrase to look for. Enclose whitespace in quotes e.g.
--phrase "small brown cows"
The shape of the image to download. Preset values are:
Thickness of border surrounding the generated gallery image in pixels. Default is 30. Enter 0 for no border.
Create the gallery in condensed mode. No padding between each thumbnail.
Put the debug log into the image sub-directory afterward. If selected, debugging output is appended to 'debug.log' in the image sub-directory. This file is always created in the temporary build directory. Great for finding out what external commands and parameters were used!
Delete the downloaded images after building the thumbnail gallery. Umm, don't specify this and
--no-gallery at the same time.
How many download failures before exiting? Default is 40. Enter 0 for unlimited (this can potentially try to download every result so only use this if you've previously had a lot of failures).
Display this help then exit.
Put your search phrases into a text file then specify the file here. googliser will download images matching each phrase in the file, ignoring any line starting with a
Only download image files larger than this many bytes. Some servers do not report a byte file-size, so these will be downloaded anyway and checked afterward (unless
--skip-no-size is specified). Default is 1,000 bytes. I've found this setting useful for skipping files sent by servers that give me HTML instead of the JPG I requested.
Only get image file URLs. Don't download any images.
Only download images with at least this many pixels. Preset values are:
qsvga(400 x 300)
vga(640 x 480)
svga(800 x 600)
xga(1024 x 768)
2mp(1600 x 1200)
4mp(2272 x 1704)
6mp(2816 x 2112)
8mp(3264 x 2448)
10mp(3648 x 2736)
12mp(4096 x 3072)
15mp(4480 x 3360)
20mp(5120 x 3840)
40mp(7216 x 5412)
70mp(9600 x 7200)
Number of images to download. Default is 25. Maximum is 1,000.
Runtime display in bland, uncoloured text.
Don't create a thumbnail gallery. Err, don't specify this and
--delete-after at the same time.
The output directory. If unspecified, the search phrase is used.
How many parallel image downloads? Default is 10. Maximum is 40.
Suppress standard display output. Error messages are still shown.
Download a single random image. Use
-n --number to set the size of the image pool to pick a random image from.
Number of download retries for each image. Default is 3. Maximum is 100.
Only get images published this far back in time. Default is any. Preset values are:
Put the URL results file into the image sub-directory afterward. If selected, the URL list will be found in 'download.links.list' in the image sub-directory. This file is always created in the temporary build directory.
Some servers do not report a byte file-size, so this parameter will ensure these image files are not downloaded. Specifying this will speed up downloading but will generate more failures.
Specify the maximum dimensions of thumbnails used in the gallery image. Width-by-height in pixels. Default is 400x400. If also using condensed-mode
-C --condensed, this setting determines the size and shape of each thumbnail. Specify like
Number of seconds before Wget gives up. Default is 5. Maximum is 600 (10 minutes).
Specify a custom title for the gallery. Default is to use the search-phrase. To create a gallery with no title, specify
--title false. Enclose whitespace in single or double-quotes according to taste. e.g.
--title 'This is what cows look like!'
Image type to download. Preset values are:
Only download image files smaller than this many bytes. Some servers do not report a byte file-size, so these will be downloaded anyway and checked afterward (unless
--skip-no-size is specified). Default is 0 (unlimited).
Usage rights. Preset values are:
reuse(labeled for reuse)
reuse-with-mod(labeled for reuse with modification)
noncomm-reuse(labeled for noncommercial reuse)
noncomm-reuse-with-mod(labeled for noncommercial reuse with modification)
Lightning mode! For those who really can't wait! Lightning mode downloads images even faster by using an optimized set of parameters: timeouts are reduced to 1 second, don't retry any download, skip any image when the server won't tell us how big it is, download up to 16 images at the same time, and don't create a gallery afterward.
$ ./googliser.sh -p "cows"
This will download the first 25 available images for the search-phrase "cows"
$ ./googliser.sh --number 250 --phrase "kittens" --parallel 12 --failures 0
This will download the first 250 available images for the search-phrase "kittens", download up to 12 images at once and ignore the failures limit.
$ ./googliser.sh --number 56 --phrase "fish" --upper-size 50000 --lower-size 2000 --failures 0 --debug
This will download the first 56 available images for the search-phrase "fish" but only if the image files are between 2KB and 50KB in size, ignore the failures limit and write a debug file.
$ ./googliser.sh -n80 -p "storm clouds" -sN --debug
This will download the first 80 available images for the phrase "storm clouds", ensure both debug and URL links files are placed in the target directory and won't create a thumbnail gallery.
0 : success!
1 : required external program unavailable.
2 : specified parameter incorrect - help shown.
3 : unable to create sub-directory for 'search-phrase'.
4 : could not get a list of search results from Google.
5 : image download aborted as failure-limit was reached or ran out of images.
6 : thumbnail gallery build failed.
7 : unable to create a temporary build directory.
I wrote this script so users don't need to obtain an API key from Google to download multiple images.
To download 1,000 images, you need to be lucky enough for Google to find 1,000 results for your search term, and for those images to be available for download. I sometimes get more failed downloads than successful downloads (depending on what I'm searching for). In practice, I've never actually had Google return 1,000 results. Best was about 986.
If identify (from ImageMagick) is installed, every downloaded file is checked to ensure that it is actually an image. Every file is renamed according to the image type determined by identify. If identify is not available, then no type-checking occurs.
Every link that cannot be downloaded, or is outside the specified byte-size range, counts as a 'failure'. A good way to see lots of failures quickly is to specify a narrow byte-size range. e.g.
--lower-size 12000 --upper-size 13000.
The failures percentage shown after download is the number of failed downloads as a percentage of the total number of image downloads attempted - this includes successful downloads. e.g. 25 images downloaded OK with 8 download failures yields a total of 33 downloads attempted. And 8 / 33 = 24%.
The final search results count is determined by adding together the number of images required (default is 25) with the number of allowable failures (default is 40). Search results initially download in groups of 100. So, for the defaults shown, the script downloads the first group of 100. Then trims it so only the first 65 results remain. Then downloads these as images. Results can be shorter though depending on other factors such as URLs returned with invalid names, Google not sending many results from the requested search, etc... The URL results list is only trimmed after dealing with these issues. The count can also change between searches as Google don't always return the same results - even for identical searches.
Only the first image of a multi-image file (like an animated GIF) will be used for its gallery image.
Thumbnail gallery building can be disabled if not required by using
--no-gallery. Why? As a test, I built a gallery from 380 images (totalling 32.6MB) which created a file of 201MB with dimensions of 8,064 x 7,876 (63.5MP). This took montage a bit over 20 minutes to render on my old Atom D510 CPU.
Typically, downloads run quite fast and then get slower as the required number of images is reached, as less parallel downloads are running (which I'll refer to as download slots). Sometimes downloads will appear to stall, as all the download slots are being held up by servers that are not responding/slow to respond or are downloading very large files. New download slots won't open up until at least one of these completes, fails or times-out. If you download a large enough number of files, all the download slots can end up like this. This is perfectly normal behaviour and the problem will sort itself out. Grab a coffee.
Sometimes, you may also see a slowdown when downloading the last image (e.g. when something like 24 out of 25 images have downloaded without issue). This leaves only one download slot available to use. However, this slot keeps encountering a series of problem links (as mentioned above) and so can take some time to get that last image as the script works it way through the links list. Grab a danish to go with that coffee.
The temporary build directory is
/tmp/googliser.PID.UNIQwhere PID is shown in the title of the script when it runs and UNIQ will be any 3 random alpha-numeric characters.
This script will need to be updated from time-to-time as Google periodically change their search results page-code. The latest copy can be found here.
- Debian - 9.8 Stretch 64b
- GNU BASH - v4.4.12
- GNU sed - v4.4
- GNU grep - v2.27
- GNU Wget - v1.18
- ImageMagick - v6.9.7-4 Q16
- Geany - v1.29
- ReText - v6.0.2
- Konsole - v16.12.0
- KDE Development Platform - v4.14.2
- QT - v5.7.1
- Find Icons - script icon
Also tested on:
- openSUSE - LEAP 42.1 64b
- Ubuntu - 18.04.1 LTS
- macOS - 10.13 High Sierra, 10.14 Mojave
- Fedora - 28 Workstation
- Mint - 19.1 Tessa XFCE
- Manjaro - 18.0.2 XFCE
Suggestions / comments / bug reports / advice (are|is) most welcome. :) email me