Occasionally, some kind of value can be accessed in multiple ways.
For example, perhaps a "User" type can be loaded not only by an "id"
but also by a "username" value. If the same user is loaded by both keys,
then it may be useful to fill both caches when a user is loaded from either source:
const userByIDLoader = newDataLoader(async ids => {
const users = awaitgenUsersByID(ids)
for (let user of users) {
usernameLoader.prime(user.username, user) // Priming in the other DataLoader.
}
return users
});
const usernameLoader = newDataLoader(async names => {
const users = awaitgenUsernames(names)
for (let user of users) {
userByIDLoader.prime(user.id, user) // Priming in the other DataLoader.
}
return users
});
Clearing Cache
In certain uncommon cases, clearing the request cache may be necessary.
The most common example when clearing the loader's cache is necessary is after a mutation or update within the same request, when a cached value could be out of date and future loads should not use any possibly cached value.
Here's a simple example using SQL UPDATE to illustrate.
// Request begins...const userLoader = newDataLoader(...)
// And a value happens to be loaded (and cached).const user = await userLoader.load(4)
// A mutation occurs, invalidating what might be in cache.awaitsqlRun('UPDATE users WHERE id=4 SET username="zuck"')
userLoader.clear(4)
// Later the value load is loaded again so the mutated data appears.const user = await userLoader.load(4)
// Request completes.
Memory consumption for long-lived DataLoaders
Custom Cache. As mentioned above, DataLoader is intended to be used as a per-request cache. Since requests are short-lived, DataLoader uses an infinitely growing Map as a memoization cache. This should not pose a problem as most requests are short-lived and the entire cache can be discarded after the request completes.
However this memoization caching strategy isn't safe when using a long-lived DataLoader, since it could consume too much memory. If using DataLoader in this way, you can provide a custom Cache instance with whatever behavior you prefer, as long as it follows the same API as Map.
The example below uses an LRU (least recently used) cache to limit total memory to hold at most 100 cached values via the lru_map npm package.
More specifically, any object that implements the methods get(), set(), delete() and clear() methods can be provided. This allows for custom Maps which implement various cache algorithms to be provided.
FAQs
> TODO: description
We found that @atomly/dataloader-sdk demonstrated a not healthy version release cadence and project activity because the last version was released a year ago.It has 2 open source maintainers collaborating on the project.
Last updated on 09 Jun 2021
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
CISA launched a new project called Vulnrichment to enrich CVEs with details that help prioritize patching and mitigation efforts, as the NVD backlog of unenriched CVEs awaiting analysis surpasses 10,000.
Socket is joining forces with CISA and other industry leaders at the RSA Conference to sign the Secure by Design pledge, committing to uphold the highest security standards in our products.
The Socket research team breaks down a sampling of malicious packages that download and execute files, among other suspicious behaviors, targeting the popular Discord platform.