An image hashing library written in Python. ImageHash supports:
- Average hashing (aHash)
- Perceptual hashing (pHash)
- Difference hashing (dHash)
- Wavelet hashing (wHash)
- HSV color hashing (colorhash)
Image hashes tell whether two images look nearly identical. This is different from cryptographic hashing algorithms (like MD5, SHA-1) where tiny changes in the image give completely different hashes. In image fingerprinting, we actually want our similar inputs to have similar output hashes as well.
The image hash algorithms (average, perceptual, difference, wavelet) analyse the image structure on luminance (without color information). The color hash algorithm analyses the color distribution and black & gray fractions (without position information).
Based on PIL/Pillow Image, numpy and scipy.fftpack (for pHash) Easy installation through pypi:
pip install imagehash
>>> from PIL import Image >>> import imagehash >>> hash = imagehash.average_hash(Image.open('test.png')) >>> print(hash) d879f8f89b1bbf >>> otherhash = imagehash.average_hash(Image.open('other.bmp')) >>> print(otherhash) ffff3720200ffff >>> print(hash == otherhash) False >>> print(hash - otherhash) 36
Each algorithm can also have its hash size adjusted (or in the case of
binbits). Increasing the hash size allows an
algorithm to store more detail in its hash, increasing its sensitivity
to changes in detail.
The demo script find_similar_images illustrates how to find similar images in a directory.
Source hosted at GitHub: https://github.com/JohannesBuchner/imagehash
To help evaluate how different hashing algorithms behave, below are a few hashes applied to two datasets. This will let you know what images an algorithm thinks are basically identical. For understanding hash distances, check out these excellent blog posts:
The first dataset is a collection of 7441 free icons on GitHub (see examples/github-urls.txt). The following pages show groups of images with the same hash (the hashing method sees them as the same).
- phash hashsize=8 clusters (with z-transform)
- dhash hashsize=8 clusters (with z-transform)
- whash-haar hashsize=8 clusters (with z-transform)
- whash-db4 hashsize=8 clusters (with z-transform)
- colorhash binbits=3 clusters
- average_hash hashsize=8 clusters (with z-transform)
The second dataset is a collection of 109259 art pieces from parismuseescollections.paris.fr/en/recherche/image-libre/.
The following pages show groups of images with the same hash (the hashing method sees them as the same).
- phash hashsize:8 clusters (with z-transform)
- dhash hashsize:8 clusters (with z-transform)
- whash-haar hashsize:8 clusters (with z-transform)
- whash-db4 hashsize:8 clusters (with z-transform)
- colorhash binbits:3 clusters
- average_hash hashsize:8 clusters (with z-transform)
You may want to adjust the hashsize or require some manhattan distance (hash1 - hash2 < threshold).
Pull requests and new features are welcome.
If you encounter a bug or have a question, please open a GitHub issue. You can also try Stack Overflow.
- 4.1: Add examples and colorhash
- 4.0: Changed binary to hex implementation, because the previous one was broken for various hash sizes. This change breaks compatibility to previously stored hashes; to convert them from the old encoding, use the "old_hex_to_hash" function.
- 3.5: Image data handling speed-up
- 3.2: whash now also handles smaller-than-hash images
- 3.0: dhash had a bug: It computed pixel differences vertically, not horizontally.
- I modified it to follow dHash. The old function is available as dhash_vertical.
- 2.0: Added whash
- 1.0: Initial ahash, dhash, phash implementations.