New Case Study:See how Anthropic automated 95% of dependency reviews with Socket.Learn More
Socket
Sign inDemoInstall
Socket

dw-cache

Package Overview
Dependencies
Maintainers
1
Versions
121
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

dw-cache

The highest performance constant complexity cache algorithm.

  • 0.0.88
  • Source
  • npm
  • Socket score

Version published
Weekly downloads
8
decreased by-74.19%
Maintainers
1
Weekly downloads
 
Created
Source

Dual Window Cache

CI

Dual window cache adaptively coordinates the ratio of LRU to LFU using the two sliding windows.

Maintenance

The source code is maintained on the next source repository.

https://github.com/falsandtru/spica

Abstract

The highest performance constant complexity cache algorithm.

Strategies

  • Dynamic partition
  • Sliding window
  • Transitive wide MRU with cyclic replacement

Properties

Generally superior and almost flawless.

  • High performance
    • High hit ratio (DS1, S3, OLTP, GLI)
      • Highest hit ratio among all the general-purpose cache algorithms.
      • Near ARC (S3, OLTP).
      • Significantly higher than ARC (DS1, GLI).
    • Low overhead (High throughput)
      • Constant time complexity overhead decreasing in linear time.
      • Use of only two lists.
    • Low latency
      • Constant time complexity.
      • No batch processing like LIRS and TinyLFU.
    • Parallel suitable
      • Separated lists are suitable for lock-free processing.
  • Efficient
    • Low memory usage
      • Constant extra space complexity.
      • Retain only keys of resident entries (No history).
    • Immediate release of evicted keys
      • Primary cache algorithm in the standard library must release memory immediately.
  • High resistance
    • Scan, loop, and burst resistance
  • Few tradeoffs
    • Not the highest hit ratio
    • Significantly small cache size can degrade hit ratio
  • Upward compatible with ARC
    • Comprehensively higher performance
  • Upward compatible with Segmented LRU
    • Totally higher performance
    • Suitable for TinyLFU
      • Better for (W-)TinyLFU's eviction algorithm.

Efficiency

Some different cache algorithms require extra memory space to retain evicted keys. Linear time complexity indicates the existence of batch processing. Note that admission algorithm doesn't work without eviction algorithm.

AlgorithmTypeTime complexity
(Worst case)
Space complexity
(Extra)
Key sizeData structures
LRUEvictConstantConstant1x1 list
DWCEvictConstantConstant1x2 lists
ARCEvictConstantLinear2x4 lists
LIRSEvictLinearLinear3-2500x2 lists
TinyLFUAdmitLinearLinear8bit * 10N * 45 arrays
W-TinyLFUAdmitLinearLinear8bit * 10N * 41 list
4 arrays

https://github.com/ben-manes/caffeine/wiki/Efficiency
https://github.com/zhongch4g/LIRS2/blob/master/src/replace_lirs_base.cc

Resistance

LIRS's burst resistance means resistance to continuous cache misses.

AlgorithmTypeScanLoopBurst
LRUEvict
DWCEvict
ARCEvict
LIRSEvict
TinyLFUAdmit
W-TinyLFUAdmit

Tradeoffs

Note that LIRS and TinyLFU are risky cache algorithms.

  • LRU
    • Low performance
    • No resistance
      • Scan access clears all entries.
  • DWC
    • Not the highest hit ratio
    • Significantly small cache size can degrade hit ratio
  • ARC
    • Middle performance
    • Inefficient
      • 2x key size.
    • High overhead
      • 4 lists.
    • Few resistance
      • No loop resistance.
  • LIRS
  • TinyLFU
    • Unreliable performance
      • Burst access degrades performance.
      • Lower hit ratio than LRU at OLTP.
      • Many major benchmarks are lacking in the paper despite the performance of TinyLFU is worse than LRU in theory.
    • Restricted delete operation
      • Bloom filters don't support delete operation.
      • Frequent delete operations degrade performance.
    • Spike latency
      • Whole reset of Bloom filters takes linear time.
    • Vulnerable algorithm
      • Burst access saturates Bloom filters.
  • W-TinyLFU
    • Restricted delete operation
      • Bloom filters don't support delete operation.
      • Frequent delete operations degrade performance.
    • Spike latency
      • Whole reset of Bloom filters takes linear time.

Hit ratio

Note that another cache algorithm sometimes changes the parameter values per workload to get a favorite result as the paper of TinyLFU has changed the window size of W-TinyLFU. All the results of DWC are measured by the same default parameter values. Graphs are approximate.

  1. Set the datasets to ./benchmark/trace (See ./benchmark/ratio.ts).
  2. Run npm i
  3. Run npm run bench
  4. Click the DEBUG button to open a debug tab.
  5. Close the previous tab.
  6. Press F12 key to open devtools.
  7. Select the console tab.

https://github.com/dgraph-io/benchmarks
https://github.com/ben-manes/caffeine/wiki/Efficiency
https://github.com/dgraph-io/ristretto
https://github.com/jedisct1/rust-arc-cache/issues/1
https://docs.google.com/spreadsheets/d/1G3deNz1gJCoXBE2IuraUSwLE7H_EMn4Sn2GU0HTpI5Y

DS1

W-TinyLFU > DWC, (LIRS) > (TinyLFU) > ARC > LRU

  • DWC is significantly better than ARC.

image

DS1 1,000,000
LRU hit ratio 3.08%
DWC hit ratio 11.34%
DWC - LRU hit ratio delta 8.25%
DWC / LRU hit ratio rate  367%

DS1 2,000,000
LRU hit ratio 10.74%
DWC hit ratio 25.76%
DWC - LRU hit ratio delta 15.02%
DWC / LRU hit ratio rate  239%

DS1 3,000,000
LRU hit ratio 18.59%
DWC hit ratio 37.93%
DWC - LRU hit ratio delta 19.34%
DWC / LRU hit ratio rate  204%

DS1 4,000,000
LRU hit ratio 20.24%
DWC hit ratio 41.59%
DWC - LRU hit ratio delta 21.34%
DWC / LRU hit ratio rate  205%

DS1 5,000,000
LRU hit ratio 21.03%
DWC hit ratio 47.33%
DWC - LRU hit ratio delta 26.30%
DWC / LRU hit ratio rate  225%

DS1 6,000,000
LRU hit ratio 33.95%
DWC hit ratio 54.65%
DWC - LRU hit ratio delta 20.70%
DWC / LRU hit ratio rate  160%

DS1 7,000,000
LRU hit ratio 38.89%
DWC hit ratio 57.77%
DWC - LRU hit ratio delta 18.87%
DWC / LRU hit ratio rate  148%

DS1 8,000,000
LRU hit ratio 43.03%
DWC hit ratio 64.64%
DWC - LRU hit ratio delta 21.61%
DWC / LRU hit ratio rate  150%

S3

W-TinyLFU > (TinyLFU) > (LIRS) > DWC, ARC > LRU

  • DWC is an approximation of ARC.

image

S3 100,000
LRU hit ratio 2.32%
DWC hit ratio 10.52%
DWC - LRU hit ratio delta 8.20%
DWC / LRU hit ratio rate  452%

S3 200,000
LRU hit ratio 4.63%
DWC hit ratio 18.87%
DWC - LRU hit ratio delta 14.24%
DWC / LRU hit ratio rate  407%

S3 300,000
LRU hit ratio 7.58%
DWC hit ratio 24.59%
DWC - LRU hit ratio delta 17.01%
DWC / LRU hit ratio rate  324%

S3 400,000
LRU hit ratio 12.03%
DWC hit ratio 29.62%
DWC - LRU hit ratio delta 17.58%
DWC / LRU hit ratio rate  246%

S3 500,000
LRU hit ratio 22.76%
DWC hit ratio 37.50%
DWC - LRU hit ratio delta 14.73%
DWC / LRU hit ratio rate  164%

S3 600,000
LRU hit ratio 34.63%
DWC hit ratio 46.28%
DWC - LRU hit ratio delta 11.65%
DWC / LRU hit ratio rate  133%

S3 700,000
LRU hit ratio 46.04%
DWC hit ratio 55.31%
DWC - LRU hit ratio delta 9.27%
DWC / LRU hit ratio rate  120%

S3 800,000
LRU hit ratio 56.59%
DWC hit ratio 63.38%
DWC - LRU hit ratio delta 6.78%
DWC / LRU hit ratio rate  111%

OLTP

W-TinyLFU > ARC, DWC > (LIRS) > LRU > (TinyLFU)

  • DWC is an approximation of ARC.

image

OLTP 250
LRU hit ratio 16.47%
DWC hit ratio 16.03%
DWC - LRU hit ratio delta -0.43%
DWC / LRU hit ratio rate  97%

OLTP 500
LRU hit ratio 23.44%
DWC hit ratio 27.14%
DWC - LRU hit ratio delta 3.69%
DWC / LRU hit ratio rate  115%

OLTP 750
LRU hit ratio 28.28%
DWC hit ratio 33.71%
DWC - LRU hit ratio delta 5.43%
DWC / LRU hit ratio rate  119%

OLTP 1,000
LRU hit ratio 32.83%
DWC hit ratio 37.57%
DWC - LRU hit ratio delta 4.74%
DWC / LRU hit ratio rate  114%

OLTP 1,250
LRU hit ratio 36.20%
DWC hit ratio 39.64%
DWC - LRU hit ratio delta 3.43%
DWC / LRU hit ratio rate  109%

OLTP 1,500
LRU hit ratio 38.69%
DWC hit ratio 41.34%
DWC - LRU hit ratio delta 2.65%
DWC / LRU hit ratio rate  106%

OLTP 1,750
LRU hit ratio 40.78%
DWC hit ratio 42.74%
DWC - LRU hit ratio delta 1.95%
DWC / LRU hit ratio rate  104%

OLTP 2,000
LRU hit ratio 42.46%
DWC hit ratio 44.11%
DWC - LRU hit ratio delta 1.65%
DWC / LRU hit ratio rate  103%

GLI

W-TinyLFU, (LIRS) > DWC > (TinyLFU) >> ARC > LRU

  • DWC is significantly better than ARC.

image

GLI 250
LRU hit ratio 0.93%
DWC hit ratio 15.25%
DWC - LRU hit ratio delta 14.32%
DWC / LRU hit ratio rate  1639%

GLI 500
LRU hit ratio 0.96%
DWC hit ratio 31.01%
DWC - LRU hit ratio delta 30.05%
DWC / LRU hit ratio rate  3217%

GLI 750
LRU hit ratio 1.16%
DWC hit ratio 41.75%
DWC - LRU hit ratio delta 40.59%
DWC / LRU hit ratio rate  3588%

GLI 1,000
LRU hit ratio 11.22%
DWC hit ratio 47.39%
DWC - LRU hit ratio delta 36.17%
DWC / LRU hit ratio rate  422%

GLI 1,250
LRU hit ratio 21.25%
DWC hit ratio 52.87%
DWC - LRU hit ratio delta 31.61%
DWC / LRU hit ratio rate  248%

GLI 1,500
LRU hit ratio 36.56%
DWC hit ratio 53.93%
DWC - LRU hit ratio delta 17.37%
DWC / LRU hit ratio rate  147%

GLI 1,750
LRU hit ratio 45.04%
DWC hit ratio 55.10%
DWC - LRU hit ratio delta 10.05%
DWC / LRU hit ratio rate  122%

GLI 2,000
LRU hit ratio 57.41%
DWC hit ratio 57.41%
DWC - LRU hit ratio delta 0.00%
DWC / LRU hit ratio rate  100%

LOOP

LOOP 100
LRU hit ratio 0.00%
DWC hit ratio 8.68%
DWC - LRU hit ratio delta 8.68%
DWC / LRU hit ratio rate  Infinity%

LOOP 250
LRU hit ratio 0.00%
DWC hit ratio 21.82%
DWC - LRU hit ratio delta 21.82%
DWC / LRU hit ratio rate  Infinity%

LOOP 500
LRU hit ratio 0.00%
DWC hit ratio 46.88%
DWC - LRU hit ratio delta 46.88%
DWC / LRU hit ratio rate  Infinity%

LOOP 750
LRU hit ratio 0.00%
DWC hit ratio 64.00%
DWC - LRU hit ratio delta 64.00%
DWC / LRU hit ratio rate  Infinity%

LOOP 1,000
LRU hit ratio 0.00%
DWC hit ratio 95.94%
DWC - LRU hit ratio delta 95.94%
DWC / LRU hit ratio rate  Infinity%

LOOP 1,250
LRU hit ratio 99.80%
DWC hit ratio 99.80%
DWC - LRU hit ratio delta 0.00%
DWC / LRU hit ratio rate  100%

Throughput

60-100% of lru-cache.

Note that the number of trials per capacity for simulation 1,000,000 is insufficient.

No result with 10,000,000 because lru-cache crushes with the next error on the next machine of GitHub Actions. It is verified that the error was thrown also when benchmarking only lru-cache. Of course it is verified that DWC works fine under the same condition.

Error: Uncaught RangeError: Map maximum size exceeded

System:
OS: Linux 5.15 Ubuntu 20.04.5 LTS (Focal Fossa)
CPU: (2) x64 Intel(R) Xeon(R) Platinum 8370C CPU @ 2.80GHz
Memory: 5.88 GB / 6.78 GB

ISCCache: lru-cache
LRUCache: spica/lru
DW-Cache: spica/cache

'ISCCache new x 10,532 ops/sec ±1.90% (111 runs sampled)'

'LRUCache new x 21,762,641 ops/sec ±1.69% (122 runs sampled)'

'DW-Cache new x 5,536,872 ops/sec ±0.16% (123 runs sampled)'

'ISCCache simulation 10 x 7,903,306 ops/sec ±2.06% (119 runs sampled)'

'LRUCache simulation 10 x 9,261,349 ops/sec ±2.59% (119 runs sampled)'

'DW-Cache simulation 10 x 5,452,576 ops/sec ±1.99% (121 runs sampled)'

'ISCCache simulation 100 x 8,120,925 ops/sec ±2.05% (119 runs sampled)'

'LRUCache simulation 100 x 8,886,277 ops/sec ±2.70% (119 runs sampled)'

'DW-Cache simulation 100 x 5,588,530 ops/sec ±1.86% (121 runs sampled)'

'ISCCache simulation 1,000 x 7,133,768 ops/sec ±2.00% (118 runs sampled)'

'LRUCache simulation 1,000 x 7,854,281 ops/sec ±2.51% (117 runs sampled)'

'DW-Cache simulation 1,000 x 5,585,856 ops/sec ±2.07% (121 runs sampled)'

'ISCCache simulation 10,000 x 6,404,151 ops/sec ±2.03% (120 runs sampled)'

'LRUCache simulation 10,000 x 7,195,591 ops/sec ±2.37% (118 runs sampled)'

'DW-Cache simulation 10,000 x 4,624,938 ops/sec ±1.74% (120 runs sampled)'

'ISCCache simulation 100,000 x 3,272,336 ops/sec ±1.57% (111 runs sampled)'

'LRUCache simulation 100,000 x 3,592,944 ops/sec ±2.29% (113 runs sampled)'

'DW-Cache simulation 100,000 x 3,468,637 ops/sec ±2.51% (115 runs sampled)'

'ISCCache simulation 1,000,000 x 1,553,743 ops/sec ±3.48% (102 runs sampled)'

'LRUCache simulation 1,000,000 x 1,802,533 ops/sec ±4.09% (105 runs sampled)'

'DW-Cache simulation 1,000,000 x 1,496,925 ops/sec ±2.55% (112 runs sampled)'
const key = random() < 0.8
  ? random() * capacity * 1 | 0
  : random() * capacity * 9 + capacity | 0;
cache.get(key) ?? cache.set(key, {});

Comprehensive evaluation

Hit ratio

ClassAlgorithms
Very highW-TinyLFU
HightDWC, (LIRS)
MiddleARC, (TinyLFU)
LowLRU

Efficiency

Extra spaceAlgorithms
ConstantLRU, DWC
Linear (< 1)W-TinyLFU > (TinyLFU)
Linear (1)ARC
Linear (> 1)(LIRS)

Resistance

ClassEffectAlgorithms
TotalHighW-TinyLFU, DWC
MostHigh(TinyLFU), (LIRS)
FewLowARC
NoneNoneLRU

Throughput

ClassAlgorithms
Bloom filter + Static list(TinyLFU)
Multiple lists (Lock-free)DWC > (LIRS) > ARC
Dynamic list + Bloom filterW-TinyLFU
Static listLRU

Latency

TimeAlgorithms
ConstantLRU, DWC, ARC
Linear (1)W-TinyLFU > (TinyLFU)
Linear (> 1)(LIRS)

Vulnerability

ClassAlgorithms
Degrade(TinyLFU)
Crush(LIRS)

API

export namespace Cache {
  export interface Options<K, V = undefined> {
    // Max entries.
    // Range: 1-
    readonly capacity?: number;
    // Window size ratio to measure hit ratios.
    // Range: 1-100
    readonly window?: number;
    // Min sample size ratio to measure hit density.
    // Range: 1-100
    readonly sample?: number;
    // Max costs.
    // Range: L-
    readonly resource?: number;
    readonly age?: number;
    readonly earlyExpiring?: boolean;
    readonly disposer?: (value: V, key: K) => void;
    readonly capture?: {
      readonly delete?: boolean;
      readonly clear?: boolean;
    };
    // Mainly for experiments.
    readonly resolution?: number;
    readonly offset?: number;
    readonly sweep?: {
      readonly threshold?: number;
      readonly window?: number;
      readonly range?: number;
      readonly shift?: number;
    };
  }
}
export class Cache<K, V = undefined> {
  constructor(capacity: number, opts?: Cache.Options<K, V>);
  constructor(opts: Cache.Options<K, V>);
  put(key: K, value: V, opts?: { size?: number; age?: number; }): boolean;
  put(this: Cache<K, undefined>, key: K, value?: V, opts?: { size?: number; age?: number; }): boolean;
  set(key: K, value: V, opts?: { size?: number; age?: number; }): this;
  set(this: Cache<K, undefined>, key: K, value?: V, opts?: { size?: number; age?: number; }): this;
  get(key: K): V | undefined;
  has(key: K): boolean;
  delete(key: K): boolean;
  clear(): void;
  resize(capacity: number, resource?: number): void;
  readonly length: number;
  readonly size: number;
  [Symbol.iterator](): Iterator<[K, V], undefined, undefined>;
}

Keywords

FAQs

Package last updated on 28 Nov 2022

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc