A lightweight, generational pseudo-LRU (Least Recently Used) cache with strict maximum size limits.
GenerationalCache maintains two internal Map objects: current and old.
-
Insertion: New items are always added to the
currentgeneration. -
Promotion: If you
getan item that exists in theoldgeneration, it is promoted to thecurrentgeneration to ensure it stays in the cache longer. -
Generation Swapping: Once the
currentgeneration reaches the boundary size ($max / 2$ ), theoldgeneration is discarded, thecurrentgeneration becomes theoldgeneration, and a new emptycurrentgeneration is created.
This "pseudo-LRU" approach avoids the overhead of updating timestamps or complex linked list pointers on every single access.
npm i @asamuzakjp/generational-cacheimport { GenerationalCache } from '@asamuzakjp/generational-cache';
// Initialize with a max capacity of 1024 items
const cache = new GenerationalCache(1024);Creates a new cache instance.
max(number): The maximum number of items the cache can hold. If the specified value is less than 4, or if an invalid value is specified, the default value of 4 will be used.
cache.size(number, read-only): Returns the total number of entries currently in the cache. Note: To optimize for write speed, this library allows temporary key duplication between generations. Therefore, this value may not always reflect the exact count of unique keys.cache.max(number): Gets or sets the maximum capacity. Note: Updating this property dynamically will invokecache.clear()to safely recalculate boundaries.
cache.get(key)Retrieves an item. If the item is found in the older generation, it is automatically promoted to the current generation to prevent it from being evicted during the next swap.- Returns: The value associated with the key, or
undefined.
- Returns: The value associated with the key, or
cache.set(key, value)Adds or updates an item. If adding this item pushes the current generation's size to the boundary threshold (max / 2), a generation swap is triggered, and the old generation is discarded.- Returns: The cache instance itself (allows chaining).
cache.has(key)Checks if a key exists in the cache (in either generation).- Returns:
trueif the key exists, otherwisefalse.
- Returns:
cache.delete(key)Removes an item from the cache.- Returns:
trueif the item existed and was removed, otherwisefalse.
- Returns:
cache.clear()Empties all items from the cache by dropping references to the internal Maps.
Benchmarks are divided into two states to simulate real-world conditions:
- Cold State: Measured with aggressive internal Garbage Collection to observe performance before full V8 TurboFan optimizations.
- Warm State: Measured after sufficient warmup, representing sustained throughput under optimal JIT compilation.
The results below reflect the sustained operations per second (ops/sec), calculated from the average latency (ns/iter). Higher values indicate better performance.
- Engine: Node.js v24.x (V8)
- Measurement: mitata.
- Comparison: LRUCache (v11.x), QuickLRU (v7.x), Mnemonist (v0.40.x)
| Scenario | State | GenerationalCache | LRUCache | QuickLRU | Mnemonist |
|---|---|---|---|---|---|
| Set | Cold | 17,733,640 ops/sec | 4,933,885 ops/sec | 13,506,212 ops/sec | 17,229,496 ops/sec |
| Warm | 23,030,861 ops/sec | 15,216,068 ops/sec | 18,175,209 ops/sec | 19,409,937 ops/sec | |
| Get | Cold | 17,717,930 ops/sec | 7,633,587 ops/sec | 13,734,377 ops/sec | 30,731,407 ops/sec |
| Warm | 21,724,961 ops/sec | 24,148,756 ops/sec | 16,385,384 ops/sec | 35,688,793 ops/sec | |
| Eviction | Cold | 16,700,066 ops/sec | 6,953,619 ops/sec | 13,285,505 ops/sec | 4,925,865 ops/sec |
| Warm | 23,148,148 ops/sec | 9,040,773 ops/sec | 16,903,313 ops/sec | 8,037,293 ops/sec |
| Scenario | State | GenerationalCache | LRUCache | QuickLRU | Mnemonist |
|---|---|---|---|---|---|
| Set | Cold | 15,987,210 ops/sec | 4,874,957 ops/sec | 11,849,745 ops/sec | 15,309,246 ops/sec |
| Warm | 19,716,088 ops/sec | 13,345,789 ops/sec | 14,755,791 ops/sec | 17,325,017 ops/sec | |
| Get | Cold | 14,994,751 ops/sec | 7,950,389 ops/sec | 11,503,508 ops/sec | 23,651,844 ops/sec |
| Warm | 17,825,311 ops/sec | 18,789,928 ops/sec | 13,838,915 ops/sec | 31,289,111 ops/sec | |
| Eviction | Cold | 16,355,904 ops/sec | 6,757,669 ops/sec | 12,074,378 ops/sec | 5,175,983 ops/sec |
| Warm | 21,982,853 ops/sec | 8,089,305 ops/sec | 15,309,246 ops/sec | 7,132,158 ops/sec |
| Scenario | State | GenerationalCache | LRUCache | QuickLRU | Mnemonist |
|---|---|---|---|---|---|
| Set | Cold | 13,679,890 ops/sec | 3,954,288 ops/sec | 8,126,777 ops/sec | 10,972,130 ops/sec |
| Warm | 20,593,080 ops/sec | 12,054,001 ops/sec | 12,995,451 ops/sec | 15,600,624 ops/sec | |
| Get | Cold | 11,918,951 ops/sec | 5,785,363 ops/sec | 9,067,827 ops/sec | 16,784,155 ops/sec |
| Warm | 16,781,339 ops/sec | 17,247,326 ops/sec | 12,733,987 ops/sec | 31,436,655 ops/sec | |
| Eviction | Cold | 13,561,160 ops/sec | 5,510,249 ops/sec | 9,642,271 ops/sec | 4,040,404 ops/sec |
| Warm | 21,128,248 ops/sec | 7,082,152 ops/sec | 13,208,294 ops/sec | 6,023,007 ops/sec |
| Metric | GenerationalCache | LRUCache | QuickLRU | Mnemonist |
|---|---|---|---|---|
| Hit Rate | 78.30% | 100.00% | 100.00% | 100.00% |
| Throughput | 10,365,916 ops/sec | 40,832,993 ops/sec | 40,950,040 ops/sec | 48,426,150 ops/sec |
| Metric | GenerationalCache | LRUCache | QuickLRU | Mnemonist |
|---|---|---|---|---|
| Hit Rate | 18.14% | 18.14% | 99.98% | 18.14% |
| Throughput | 13,232,764 ops/sec | 9,535,616 ops/sec | 10,417,752 ops/sec | 7,590,709 ops/sec |
- High Eviction Efficiency:
GenerationalCachedemonstrates strong throughput during high-turnover workloads, maintaining a performance margin compared to standard LRU designs in large-scale eviction scenarios. - Predictable Scalability: While other libraries may experience performance degradation as cache size increases,
GenerationalCachemaintains consistent throughput due to its generational swap mechanism. - Balanced Read/Write: It provides stable and competitive performance across all basic operations (
get,set), making it suitable for both read-heavy and write-heavy environments. - Trade-offs: In cyclic access patterns where the working set is greater than
max / 2but smaller thanmax,GenerationalCachewill experience frequent generation swaps and cache misses. To maximize the performance benefits ofGenerationalCache, it is often better to keep themaxsize small enough to allow some evictions, rather than trying to fit the entire working set (See 4 & 5).
MIT