This is a library providing both a higher-order function and a decorator to cache the result of a function/method if given conditions are met.
You can install it by using the following command:
npm install @jointly/cache-candidate
In this scenario, we want to cache the result of the function if the same parameters are passed 3 times in the last 30 seconds, but we want to keep the cache record for 60 seconds.
import { cacheCandidate } from '@jointly/cache-candidate';
function getUsers(filters = {}) {
return db.query('SELECT * FROM users WHERE ?', filters);
}
const cachedGetUsers = cacheCandidate(getUsers, {
requestsThreshold: 3,
ttl: 60000,
timeFrame: 30000
});
await cachedGetUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3 in the last 30 seconds
await cachedGetUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3 in the last 30 seconds
await cachedGetUsers({ name: 'John' }); // <-- This WILL be cached, because the requestsThreshold is 3 in the last 30 seconds!
await cachedGetUsers({ name: 'Jack' }); // <-- This won't be cached, because parameters are different
await sleep(61000); // <-- This will wait enough time (ttl) for the cache record to expire
await cachedGetUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3 in the last 30 seconds
In this scenario, we want to cache the result of the function if the same parameters are passed 3 times in the last 45 seconds, but we want to keep the cache record for 30 seconds. This example could reflect a scenario in which you are paying for the cache storage, yet you want to cache the same result again if the same parameters are passed 3 times in a short period of time.
import { cacheCandidate } from '@jointly/cache-candidate';
function getUsers(filters = {}) {
return db.query('SELECT * FROM users WHERE ?', filters);
}
const cachedGetUsers = cacheCandidate(getUsers, {
requestsThreshold: 3,
ttl: 30000,
timeFrame: 45000 // <-- Notice the different timeFrame, higher than cache ttl
});
await cachedGetUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3 in the last 45 seconds
await cachedGetUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3 in the last 45 seconds
await cachedGetUsers({ name: 'John' }); // <-- This WILL be cached, because the requestsThreshold is 3 in the last 45 seconds!
await cachedGetUsers({ name: 'Jack' }); // <-- This won't be cached, because parameters are different
await sleep(31000); // <-- This will wait enough time (ttl) for the cache record to expire
await cachedGetUsers({ name: 'John' }); // <-- This WILL be cached, because the requestsThreshold is 3 in the last 45 seconds!
You can also pass a candidate function to check if the conditions are met.
import { cacheCandidate } from '@jointly/cache-candidate';
function getUsers(filters = {}) {
return db.query('SELECT * FROM users WHERE ?', filters);
}
const cachedGetUsers = cacheCandidate(getUsers, {
ttl: 30000,
candidateFunction: ({ timeFrameCacheRecords, options, args }) => args[0].name === 'John',
});
await cachedGetUsers({ name: 'John' }); // <-- This will be cached, because the candidateFunction returns true
await cachedGetUsers({ name: 'John' }); // <-- This will return the cached value
await cachedGetUsers({ name: 'Jack' }); // <-- This won't be cached, because the candidateFunction returns false
await sleep(31000); // <-- This will invalidate the cache because of the ttl
The library exposes the cacheCandidate
function which accepts the function to be cached as the first argument and the options as the second argument.
The returned function is an async function which returns a Promise fulfilled with the cached value if the method has already been called with the same arguments and/or the conditions are met.
The options available are:
-
ttl
(optional): The time to live of the cache record in milliseconds. Default:600000
(10 minutes). -
timeFrame
(optional): The timeframe considered for the condition checks. Default:30000
(30 seconds).
Consider the timeFrame asthe execution history of the last X milliseconds
. This timeframe collects information about the function's execution time. For example, if you set the timeFrame to 30000 (30 seconds), the library will check the execution history of the last 30 seconds.
This means that if you set therequestsThreshold
(explained below) to 3, the function will be cached only if the same parameters are passed 3 times in the last 30 seconds. -
candidateFunction
(optional): The function to be called to check if the conditions are met. If not passed, this criteria will be ignored.
The candidateFunction receives an object with the following properties:options
: The options passed to thecacheCandidate
function.executionTime
: The execution time of the current function execution in milliseconds.args
: The arguments passed to the current function.timeFrameCacheRecords
: The cache records of the lasttimeFrame
milliseconds.
-
millisecondThreshold
(optional): The threshold in milliseconds to be considered for the condition checks. If not passed, this criteria will be ignored. -
requestsThreshold
(optional): The number of requests to be considered for the condition checks. Default:1
. -
expirationMode
(optional): The expiration mode to use. Default:default
.default
: The cache-candidate will be responsible for generating timeouts that, when reached, will call the provided delete method of the cache adapter.timeout-only
: The cache-candidate will be responsible for generating timeouts that, when reached, will not call the provided delete method of the cache adapter. This means your cache adapter must have a mechanism to delete the cache record when the timeout is reached (Ex. Redis EX option). Plugins and events will be called as usual.eject
: The cache-candidate will not be responsible for generating timeouts nor calling the delete method. This means your cache adapter must have a mechanism to delete the cache record when the timeout is reached (Ex. Redis EX option). Be aware plugins and events will not be called.
-
keepAlive
(optional): Iftrue
, the cache record will be kept alive at every request. Default:false
. -
cache
(optional): The cache adapter to be used. Defaults toan in-memory cache based on Maps, but with Promises
.
Available adapters are:makeRedisCache
: A cache adapter based on Redis. Receives a Redis client as the first and only argument.
-
fetchingMode
(optional): The fetching mode to use.
Available modes are:default
: The default behaviour, if an entry is found in the cache, it will be returned.stale-while-revalidate
: If an entry is found in the cache, it will be returned. Otherwise, a stale result will be returned and a new call to the cache-candidate will be made in background. Be aware this mechanism doesn't grant the result will be cached, depending on your threshold configuration. If, for example, you set arequestsThreshold
of 3, the revalidation will only call the function once, so it won't be cached.
-
events
(optional): Listener functions to be called at specific steps of the process.
Available events are:onCacheHit
: Called when the cache entry is hit.onCacheSet
: Called when the cache entry is set.onCacheDelete
: Called when the cache entry is deleted.onBeforeFunctionExecution
: Called before the function execution.onAfterFunctionExecution
: Called after the function execution. Every event receives an object containing thekey
property, which is the key used for the cache.
TheonAfterFunctionExecution
event also receives theexecutionTime
property, which is the execution time of the current function in milliseconds.
-
plugins
(optional): An array of plugins to be used. Default:[]
.
Please, refer to the @jointly/cache-candidate-plugin-base package for more information.
The decorator expects to receive the options as the first argument and works exactly as the higher-order function.
import { CacheCandidate } from '@jointly/cache-candidate';
class MyClass {
@CacheCandidate({
requestsThreshold: 3,
ttl: 30000,
})
async getUsers(filters = {}) {
return db.query('SELECT * FROM users WHERE ?', filters);
}
}
const myInstance = new MyClass();
await myInstance.getUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3
await myInstance.getUsers({ name: 'John' }); // <-- This won't be cached, because the requestsThreshold is 3
await myInstance.getUsers({ name: 'John' }); // <-- This WILL be cached, because the requestsThreshold is 3!
The conditions are, within the given timeFrame
:
- If a
candidateFunction
is provided, it returnstrue
at least once.
The candidateFunction ignores all the other conditions. - If a
millisecondThreshold
is provided, the function execution time passed such threshold at leastrequestsThreshold
times. - If only a
requestsThreshold
is provided (default), the function is called at leastrequestsThreshold
times.
The library supports plugins to extend its functionality.
Please, refer to the @jointly/cache-candidate-plugin-base package for the documentation on how to create a plugin.
- @jointly/cache-candidate-plugin-dependency-keys: A plugin allowing to define dependency keys when setting the cache record.
This provides a mechanism to delete one or more cache records when a dependency key is invalidated. - @jointly/cache-candidate-plugin-invalidate-function: A plugin providing an invalidation mechanism under specific conditions.
- The higher-order function and the decorator only work with async functions.
Please, refer to the Considerations on synchronous functions section for more information.
The library prevents the cache stampede problem by using a Map
called runningQueries
which saves the promise of the function call.
If multiple calls are made to the same function with the same arguments, the function will be called only once and the other calls will wait for the Promise to finish.
The runningQueries
Map will be cleaned after the function execution is finished.
The onCacheHit
event will be called also when the running query is returned.
The library doesn't consider the correct execution of the given cache functions.
It isn't the library's responsibility to check if the cache functions are working properly.
The only consideration done is based on the fact that the set
method will eventually fulfill or reject the Promise as it uses the .finally
Promise method to delete the key from the runningQueries
Map and set the timeout to clean the cache record.
If the given function is synchronous, the library could not work as expected.
The reason is that the library internally transforms the function to an asynchronous one, so executing the same function multiple times during the same Event Loop tick will prevent the cache from setting the value in time, thus not working as expected.
The expected result will still be achieved, but please consider the multiple cache set operations during development.
The cache key is composed based on the following criteria, allowing multiple files to export the same class name with the same method name / the same function names without conflicts.
- The arguments passed to the method. (JSON.stringify)
uniqueIdentifier
: A uniqid generated to allow multiple files to contain the same function name.
- The method name.
uniqueIdentifier
: A uniqid generated to allow multiple files to contain the same class with the same method.instanceIdentifier
: A uniqid generated for each instance of the class. It uses the instance properties to generate the id. (JSON.stringify)- The arguments passed to the method. (JSON.stringify)
An overhead benchmark is available launching the npm run benchmark
command.
The benchmark will run:
- A simple function without cache-candidate.
- A cache-candidate wrap of the same function of the first test with an unreachable requestsThreshold.
- A cache-candidate wrap of the same function of the first test with a candidateFunction that always returns false.
The results on a MacBook Pro M1 2021 with 16GB of RAM are:
cacheCandidate OFF - normalFunction x 47.48 ops/sec ±0.48% (74 runs sampled)
cacheCandidate ON - Request Threshold unreachable x 46.86 ops/sec ±0.76% (73 runs sampled)
cacheCandidate ON - Candidate Function set to false x 46.75 ops/sec ±0.47% (72 runs sampled)
The results show that the library has a very low overhead, even when the cache is not used.
Our suggestion therefore is to use the library on every function that could benefit from an hypothetical caching.
Please, refer to the CONTRIBUTING.md file for more information.