As already written in the description, this program downloads specified URLs, parses HTML and writes the parsed information to an atom-feed.
Thus, you do not need any account or login to be up to date what is happening on these sites. You just need a RSS-reader and an Internet-connection.
The only dependency you need to run this script is Python 3.x. If it is not installed already, you can simply do this for a Debian-like OS by running
sudo apt-get install python3
Simply download (or clone) this repository. You're done!
To run the program once, you can simply run
python Main.py
in a terminal.
For periodically usage – IMHO the main task – you can run it with the help of cron, runwhen, systemd or equivalent software. See their documentation on how to use it, please.
For further configuration – f.e. which sites should be parsed – see the file config/html2rss.cfg.default.
This file contains comments that should document the file itself.
If the download of a page fails with f.e. a 503 “Service Unavailable”, an error message with the pageurl and status-code will be directly printed to stdout. Additionally, the page will be skipped in this run.
To run the (basic) tests simply execute
python -m unittest