pwr (paced web reader) is a script and terminal-centric workflow I use for keeping up to date with various sources online, shared on the off chance it's useful to you too.
See its page on Muxup.com for more information about why it exists and how to use it.
- Ships with an example list of sources and "fetchers" (functions that will return extracted URLs and titles). This includes examples of using BeautifulSoup to extract from HTML in cases where RSS feeds are not present or are insufficient.
- Data is stored in JSON in
$XDG_DATA_HOME/.local/share/pwr
. It does assume Python 3.6+ dictionary semantics of entries being maintained in insertion order. - Supporting fetching data from multiple sources in parallel is an obvious extension. It simply hasn't been slow enough so far to make it worth prioritising.
- Caching of RSS feeds and similar isn't implemented. This is fine if you're running it once every day or so as intended, but if you're looking to run it more regularly you should add improvements in that area.
- pwr is a quick script that scratches an itch. It's definitely not the world's most robust or clean Python code.