Skip to content
Tells you which "words" can be written as a publicly available domain
Python
Branch: master
Clone or download
Latest commit 77169b8 Sep 5, 2019
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
src Deterministic order, easier usage as library Sep 5, 2019
.gitignore Add script and results Sep 5, 2019
COPYRIGHT.txt Initial commit: import ressources Sep 5, 2019
LICENSE Initial commit: import ressources Sep 5, 2019
README.md Add README Sep 5, 2019
results.txt Deterministic order, easier usage as library Sep 5, 2019

README.md

domain-word

Tells you which "words" can be written as a publicly available domain.

Domain names that end up being a word after removing the dots are cool. I guess we all remember Del.icio.us, and some may know bullsh.it.

But English has many words, and so do all the other languages. And there are many TLDs and public suffixes available. This means many domain "words" aren't taken (or even hogged) yet, and also that it's not easy to check whether a specific word is taken (because there could be many suffixes).

This project allows you to quckly (and dirtily) check whether (and how) a specific word can still be acquired. There is a surprising number still available: ev.il, mourni.ng, sna.ps, enor.mo.us, etc.

Table of Contents

Install

No need to install it. There are no dependencies.

Usage

For American English

For convenience, the result for a very exhaustive American English wordlist is already available in results.txt.

The output format is quite easy:

bathhouses: ['ses', 'es']

This means that ses is a public suffix and/or a TLD, and so is es.

For your own wordlist

If you want to substitude your own wordlist, you can replace src/american-english. Or make the code point to your own file.

As a library

You can even use it as a library! The easiest way is probably to call compute_results(wordlist_filename, source_dir=None) and process the results to your own liking.

Performance

The American English wordlist seems to be a good benchmark:

The code reads two domain lists with a combined length of over 14,000 lines, reads the wordlist of over 100,000 words, and outputs all possible combinations in under 3 seconds (2.7 seconds on my machine).

I'm sure this can be done faster, but I don't see any point in speeding this up.

Some interesting domains

Copyright / License

I license this project and my code under the MIT License. So do whatever you want with it.

Copyright and license information for the third-party sources can be found in COPYRIGHT.txt.

Contribute

Feel free to dive in! Open an issue or submit PRs.

You can’t perform that action at this time.