Want news? Just
Krill are filter feeders. True to its namesake,
krill is beautifully minimal.
krill is extremely easy to set up and use, and runs anywhere Python runs.
krill is a refreshingly different way of consuming news
krill is the hacker's way of keeping up with the world.
pip install krill
either as a superuser or from a virtualenv environment.
Of course, you can also download the script directly from this repository, in which case you will need to install the dependencies Beautiful Soup (what a library!
krill [-h] [-s URL [URL ...]] [-S FILE] [-f REGEX [REGEX ...]] [-F FILE] [-u SECONDS] -s URL [URL ...], --sources URL [URL ...] URLs to pull data from -S FILE, --sources-file FILE file from which to load source URLs (OPML format assumed if filename ends with ".opml") -f REGEX [REGEX ...], --filters REGEX [REGEX ...] patterns used to select feed items to print -F FILE, --filters-file FILE file from which to load filter patterns -u SECONDS, --update-interval SECONDS time between successive feed updates (default: 300 seconds, 0 for single pull only)
krill -s "https://twitter.com/nasa" -f "new ?horizons"
will follow NASA's
krill automatically determines whether to treat a web document as a Twitter or an XML feed. If multiple sources and/or filters are loaded from a file with the
-F tags, each must be on a separate line (except if the sources file uses the OPML format, in which case all
xmlUrl attributes are loaded). Empty lines and lines starting with
# (comments) are ignored.
Any URL format accepted by the Requests library is supported. In particular, feeds requiring (basic) HTTP authentication can be accessed by supplying credentials in the URL string with
Inline and file specifications may be combined freely. If more than one filter is given, items matching any of the filters are printed. If no filter is given, all items are printed.
Copyright © 2015 Philipp Emanuel Weidmann (email@example.com)
Released under the terms of the GNU General Public License, version 3