Rss feed parser, with collector on Raspberry
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Failed to load latest commit information.

Rss Browser A.K.A. Czytajka

Based on Elasticsearch browser that will help me to

  • find remote job
  • play with ES queries

I want also to try Stempel, Polish extension for Elasticsearch.



You can use Raspberry or other docker-ready machine to collect data. This will require using image from app folder. On your docker-ready machine just clone the repo and build the image using script. Then run the existing image using f.e. script, linking the volumes. By default output CSV will be written to /opt/rssparser/data folder. Each execution of script will make CSV bigger, filling it RSS feed entries. As for the beginning the rss will be hardcoded to URL. You can plan the execution of script periodically, f.e. putting as cron expression.


To index your files you can use prepared Logstash and Elasticsearch images in their respective folders. There is also python based loader made for this purpose. First of all, download CSV file from the machine collecting data. Then you can index files into elasticsearch. Elastisearch comes with the prepared image, with polish Stempel included in it. You can run the Elasticsearch+Kibana using docker-compose.yml script.

docker-compose up


Following there is a similar query you can run against the index after successful indexing:

POST infopraca_rss/offer/_search
	"query": {
		"bool": {
			"must": [
				{"match": {"text": "elasticsearch"}}
			"should": [
				{"match": {"text": "java"}},
				{"match": {"text": "python"}}
	"highlight": {
		"fields": {
			"text": {}