From 36a54167a7d8d7296a0beef76122cc6457ef79fb Mon Sep 17 00:00:00 2001 From: Jannis R Date: Tue, 13 Feb 2024 20:51:39 +0100 Subject: [PATCH] =?UTF-8?q?minor=20tweak=20=F0=9F=93=9D;=208.2.6?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- package.json | 2 +- readme.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/package.json b/package.json index 349ec36..7b17522 100644 --- a/package.json +++ b/package.json @@ -1,7 +1,7 @@ { "name": "hafas-gtfs-rt-feed", "description": "Generate a GTFS Realtime feed by monitoring a HAFAS endpoint.", - "version": "8.2.5", + "version": "8.2.6", "main": "index.js", "bin": { "monitor-hafas": "monitor.js", diff --git a/readme.md b/readme.md index bcedd01..ca8de1c 100644 --- a/readme.md +++ b/readme.md @@ -111,7 +111,7 @@ module.exports = hafasClient `match-with-gtfs` – `hafas-gtfs-rt-feed`'s 2nd processing step – needs a pre-populated matching database in order to match data fetched from HAFAS against the [GTFS Static/Schedule](https://gtfs.org/schedule/) data; It uses [`gtfs-via-postgres`](https://npmjs.com/package/gtfs-via-postgres) and [`match-gtfs-rt-to-gtfs`](https://npmjs.com/package/match-gtfs-rt-to-gtfs) underneath to do this matching. -First, we're going to use [`gtfs-via-postgres`](https://npmjs.com/package/gtfs-via-postgres)'s `gtfs-to-sql` command-line tool to import our GTFS data into PostgreSQL. +First, we're going to use [`gtfs-via-postgres`](https://npmjs.com/package/gtfs-via-postgres)'s `gtfs-to-sql` command-line tool to import our GTFS data into PostgreSQL. *Note:* Make sure you have an up-to-date [static GTFS](https://gtfs.org/schedule/) dataset, unzipped into individual `.txt` files.