Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update dataloader to the latest version 馃殌 #78

Open
wants to merge 2 commits into
base: master
Choose a base branch
from

Conversation

greenkeeper[bot]
Copy link

@greenkeeper greenkeeper bot commented Nov 18, 2019

The dependency dataloader was updated from 1.4.0 to 2.0.0.

This version is not covered by your current version range.

If you don鈥檛 accept this pull request, your project will work just like it did before. However, you might be missing out on a bunch of new features, fixes and/or performance improvements from the dependency update.


Publisher: leebyron
License: This package鈥檚 license has changed from BSD-3-Clause to MIT in this release 馃

Release Notes for v2.0.0

This is the first release since becoming part of the GraphQL Foundation and the most significant since the initial release over four years ago. Read more about the history of the project and this release in the blog post.

Breaking:

  • #216: .loadMany() now returns an array which may contain Error if one of the requested keys failed.

    Previously .loadMany() was exactly the same as calling Promise.all() on multiple .load() calls. While syntactically a minor convenience, this wasn't particularly useful over what could be done with Promise.all directly and if one key failed, it meant the entire call to .loadMany() would fail. As of this version, .loadMany() can now return a mix of values and Error instances in the case that some keys failed, but the Promise it returns will never be rejected. This is similar to the behavior of the new Promise.allSettled method in the upcoming version of JavaScript.

    This will break any code which relied on .loadMany(). To support this change, either ensure the each item in the result of .loadMany() are checked against instanceof Error or replace calls like loader.loadMany([k1, k2]) with Promise.all([loader.load(k1), loader.load(k2)).

  • #220: The timing of calls to batchLoadFn when { batch: false } has changed to the end of the run-loop tick.

    Previously when batching was disabled the batchLoadFn would be called immediately when .load() is called. This differed from the batchLoadFn being called at the end of the tick of the run-loop for when batching was enabled. This timing difference could lead to subtle race conditions for code which dynamically toggled batching on or off. As a simplification, the batchLoadFn is now always called at the end of the run-loop tick regardless of whether batching is disabled.

    Hopefully this will not break your code. It could cause issues for any code which relied on this synchronous call to batchLoadFn for loaders where batching was disabled.

  • #222: Promises for cached values now wait to resolve until the rest of the batch resolves.

    Previously when .load() encountered a cached value it would return an already resolved (or rejected) Promise. However when additional dependent loads happened after these, the difference in time between the cache hit value resolving and the cache miss value resolving would result in additional unnecessary network requests. As of this version when .load() encounters a cached value it returns a Promise which waits to resolve until the call to batchLoadFn also resolves. This should result in better whole-program performance and is the most significant conceptual change and improvement. This is actually not a new innovation but a correction to match the original behavior of Facebook's "Loader" from 2010 this library is inspired by.

    This changes the timing of when Promises are resolved and thus could introduce subtle behavioral change in your code, especially if your code is prone to race conditions. Please test carefully.

    This also means each return of .load() is a new Promise instance. Where prior versions returned the same Promise instance for cached results, this version does not. This may break code which uses the returned Promise as a memoization key or in some other way assumed reference equality.

  • #226: The names of private class variables have changed

    This really shouldn't break your code because you definitely don't reach into class private variables, right? I just figured it would be something you'd like to know, you know... just in case.

New:

  • #176 #209: MIT licensed (no longer BSD+Patents) and copyrights moved from Facebook to the GraphQL Foundation
  • #182: The DataLoader instance is now available as this in batchLoadFn
  • #228: Support for custom batch scheduling functions

    The dirty secret of DataLoader is that most of it is quite boring. The interesting bit is the batch scheduling function which takes advantage of Node.js's unique run-loop scheduler to acheive automatic batching without any additional latency. However since its release, ports to other languages have found this bit to be not be easily replicated and have either replaced it with something conceptually simpler (like manual dispatch) or with a scheduler custom fit to a GraphQL execution engine. These are interesting innovations which deserve ground for experimentation in this original library as well.

    Via batchScheduleFn, you can now provide a custom batch scheduling function and experiment with manual dispatch, added latency dispatch, or any other behavior which might work best for your application.

Types:

  • #145: Improved TypeScript/Flow types for custom cacheKeyFn and cacheMap
  • #146: TypeScript types allow batchLoadFn to return a PromiseLike, supporting use of bluebird
  • #214 #219: TypeScript/Flow types allow batchLoadFn to return ArrayLike, supporting returning read-only arrays
  • #168: Flow types now use strict mode, allowing safe import into other strict mode code
  • #217: Fixed an issue where TypeScript/Flow would incorrectly report an error when providing an Error to .prime()

Fixes:

  • #215: Fixed an issue where a cache could still consume memory, even when caching was disabled
  • #223: Fixed an issue where providing an Error to .prime() could incorrectly cause an unhandled promise rejection warning

Documentation:

  • Added references to a ton more ports of DataLoader into other languages (keep 'em coming!)
  • #213: All examples have been updated to latest JavaScript (preferring async/await over Promise chaining)
  • Improved documentation for custom cacheMap along with an LRU example.
  • Improved documentation for using higher-order functions on batchLoadFn.
  • Improved documentation for converting Map results to Array results in batchLoadFn.
Commits

The new version differs by 48 commits.

  • 0c05d28 2.0.0
  • 6592183 Support custom schedulers (#228)
  • 3b2192c Add disclaimer for video walkthrough
  • 01c829d Resolve all options during construction (#226)
  • cde0cd8 Update issue templates (#227)
  • 06c403b [BREAKING] Resolve cached values after batch dispatch (#222)
  • b5d7bf5 Refactor batching logic (#220)
  • a3dd591 [BREAKING] loadMany() returns individual Error instead of rejecting promise. (#216)
  • 2f7af56 Add example in documentation for converting object results to Array results.
  • 923a75f Add example for freezing results with a HOF
  • 200b522 Merge both custom cache documentations together
  • 23bd362 Add documentation for custom cacheMap field.
  • 29811db [FIX] Fix case where priming a cache with an Error results in UnhandledPromiseRejection (#223)
  • 4212c9e Add tests for behavior when process.nextTick does not exist (#221)
  • 44977c0 [FIX] Returns the keys type of the batch fn to ReadonlyArray (#219)

There are 48 commits in total.

See the full diff


FAQ and help

There is a collection of frequently asked questions. If those don鈥檛 help, you can always ask the humans behind Greenkeeper.


Your Greenkeeper bot 馃尨

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

0 participants