Split Streams3 style
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
.gitignore Basic split working. Mar 11, 2014
.travis.yml Added Node 11 to .travis.yml Dec 14, 2018
LICENSE Updated copyright year to 2018 Feb 2, 2018
README.md Add skipOverflow option Dec 12, 2018
bench.js Add binary-split into bench May 15, 2018
index.js Add skipOverflow option Dec 12, 2018
package.json Bumped v3.1.0. Dec 14, 2018
test.js Add skipOverflow option Dec 12, 2018

README.md

Split2(matcher, mapper, options)

Greenkeeper badge

build status

Break up a stream and reassemble it so that each line is a chunk. split2 is inspired by @dominictarr split module, and it is totally API compatible with it. However, it is based on Node.js core Transform via readable-stream

matcher may be a String, or a RegExp. Example, read every line in a file ...

  fs.createReadStream(file)
    .pipe(split2())
    .on('data', function (line) {
      //each chunk now is a separate line!
    })

split takes the same arguments as string.split except it defaults to '/\r?\n/', and the optional limit paremeter is ignored. String#split

split takes an optional options object on it's third argument, which is directly passed as a Transform option.

Additionally, the .maxLength and .skipOverflow options are implemented, which set limits on the internal buffer size and the stream's behavior when the limit is exceeded. There is no limit unless maxLength is set. When the internal buffer size exceeds maxLength, the stream emits an error by default. You may also set skipOverflow to true to suppress the error and instead skip past any lines that cause the internal buffer to exceed maxLength.

Calling .destroy will make the stream emit close. Use this to perform cleanup logic

var splitFile = function(filename) {
  var file = fs.createReadStream(filename)

  return file
    .pipe(split2())
    .on('close', function() {
      // destroy the file stream in case the split stream was destroyed
      file.destroy()
    })
}

var stream = splitFile('my-file.txt')

stream.destroy() // will destroy the input file stream

NDJ - Newline Delimited Json

split2 accepts a function which transforms each line.

fs.createReadStream(file)
  .pipe(split2(JSON.parse))
  .on('data', function (obj) {
    //each chunk now is a js object
  })

However, in @dominictarr split the mapper is wrapped in a try-catch, while here it is not: if your parsing logic can throw, wrap it yourself.

Benchmark

$ node bench.js
benchSplit*10000: 1484.983ms
benchBinarySplit*10000: 1484.080ms
benchSplit*10000: 1407.334ms
benchBinarySplit*10000: 1500.281ms

Benchmark taken on Node 8.11.3, on a Macbook i5 2018.

License

Copyright (c) 2014-2018, Matteo Collina hello@matteocollina.com

Permission to use, copy, modify, and/or distribute this software for any purpose with or without fee is hereby granted, provided that the above copyright notice and this permission notice appear in all copies.

THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.