Skip to content

sergeyampo/JSON-streaming-parser-to-db

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 

Repository files navigation

JSON-streaming-parser-to-db

This is a simple example of using EventEmitters, Streams, json-stream, axios and minter.network for extracting parsing and calculating data on demand and pushing it to the database.

Installation:

git clone https://github.com/sergeyampo/JSON-streaming-parser-to-db/
cd JSON-streaming* && npm i pg stream-json stream-chain axios decimal.js-light config --save-dev
npm start

Configuration:

You should configure json file in a ./config/db.json for correct database connection. For example, it's local PostgreSQL database for me.

What happens:

Diagram

Isn't it easier to JSON.parse():

No it's not. It's one of the most common mistake in blocking Event Loop in Node by using JSON.parse because it's synchronous function and it wouldn't get into the Worker Pool or in C++ Api's. Every client of the server should be await of parse function to be completed.

About

An example of how we can use streams, events, JSON and database to get JSON from server, extracting and calculating data on demand and pushing it to the database

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published