Tells you which "words" can be written as a publicly available domain.
But English has many words, and so do all the other languages. And there are many TLDs and public suffixes available. This means many domain "words" aren't taken (or even hogged) yet, and also that it's not easy to check whether a specific word is taken (because there could be many suffixes).
Table of Contents
No need to install it. There are no dependencies.
For American English
For convenience, the result for a very exhaustive American English wordlist is already available in
The output format is quite easy:
bathhouses: ['ses', 'es']
This means that
ses is a public suffix and/or a TLD, and so is
For your own wordlist
If you want to substitude your own wordlist,
you can replace
Or make the code point to your own file.
As a library
You can even use it as a library!
The easiest way is probably to call
compute_results(wordlist_filename, source_dir=None) and process the results to your own liking.
The American English wordlist seems to be a good benchmark:
The code reads two domain lists with a combined length of over 14,000 lines, reads the wordlist of over 100,000 words, and outputs all possible combinations in under 3 seconds (2.7 seconds on my machine).
I'm sure this can be done faster, but I don't see any point in speeding this up.
Some interesting domains
Copyright / License
I license this project and my code under the MIT License. So do whatever you want with it.
Copyright and license information for the third-party sources
can be found in
Feel free to dive in! Open an issue or submit PRs.