Switch branches/tags
Find file Copy path
Fetching contributors…
Cannot retrieve contributors at this time
37 lines (29 sloc) 1.22 KB


Anemone is a web spider framework that can spider a domain and collect useful information about the pages it visits. It is versatile, allowing you to write your own specialized spider tasks quickly and easily.

See for more information.


  • Multi-threaded design for high performance

  • Tracks 301 HTTP redirects

  • Built-in BFS algorithm for determining page depth

  • Allows exclusion of URLs based on regular expressions

  • Choose the links to follow on each page with focus_crawl()

  • HTTPS support

  • Records response time for each page

  • CLI program can list all pages in a domain, calculate page depths, and more

  • Obey robots.txt

  • In-memory or persistent storage of pages during crawl, using TokyoCabinet, MongoDB, or Redis


See the scripts under the lib/anemone/cli directory for examples of several useful Anemone tasks.


  • nokogiri

  • robots


To test and develop this gem, additional requirements are:

  • rspec

  • fakeweb

  • tokyocabinet

  • mongo

  • redis

You will need to have Tokyo Cabinet, MongoDB, and Redis installed on your system and running.