Not for use in production
A xml-crawler, which crawls specified site-feeds (in sites file in root) to save all posts on this websites with the content. After one hour, it will visit the posts again and check for changes. This changes will be saved in the database as a diff from previous snapshot.
All sites will be visited by its own thread, so it is faster then single-threaded.
- a lot more (diffhelper)
Type python WebMining.py help into console to get the help-message for usage.
The script is not well documented. You have to add your feed-urls into the sites-file in root to crawl them. They will be added into the database at next program start.
This script runs well in a tmux window, so you can keep track of multiple website-feeds on a raspberry pi.