Lightweight middleware function that automatically caches SQL query results for increased performance.
Analyze & Visualize »
Report Bug
·
Request Feature
Table of Contents
Before the development of this NPM package to automatically cache query results from a relational database, a web application was made for developers to analyze and visualize the performance of SQL queries. With these insights into the performance of an application's backend queries, developers can implement data-backed thresholds with the queryflow.js NPM package, such that queries slower than the set threshold will be stored in a cache database. This increases performance of applications by reducing each query's time by up to 94%.
This package assumes that a Redis instance has been set up alongside a primary relational database.
For more information about setting up a Redis database, visit their docs.
To get started:
npm install queryflow.js
In the code below, the primary relational database model, Redis model and npm package are imported into a backend controller. Here, the developer defines the SQL query string, an array of values which will be bound to the query string, the threshold time in seconds for caching the results from the primary database, and the TTL of the data stored in the cache database. The autoCache method is invoked asynchronously within the controller function. The results can be assigned, and pass on through the response cycle.
import db from "../models/ourDBModel.mjs";
import redisModel from "../models/redisModel.mjs";
import QueryFlow from "queryflow.js";
const exampleController = {};
exampleController.example = async (req, res, next) => {
const queryString =
"SELECT * FROM users WHERE firstname = $1 AND lastname = $2";
const { firstName, lastName } = req.body;
const values = [firstName, lastName];
const threshold = 3000; //Milliseconds
const TTL = 30; //Seconds
try {
const result = await QueryFlow.autoCache({
redisModel,
db,
queryString,
values,
threshold, //OPTIONAL: Default value 3 seconds.
TTL, //OPTIONAL: Default value 30 minutes.
log: true, //OPTIONAL: Default value false. Turn console logs on and off.
instanceLatency: true, //OPTIONAL: Default value false. Switches measurement of time from measuring total latency to measuring total time within the primary database itself.
});
res.locals.data = result.rows;
return next();
} catch (error) {
return next({
log: "Error handler caught error in middleware",
status: 500,
message: "Error handler caught error in middleware",
});
}
};
export default exampleController;
- redisModel: Redis client.
- db: Primary relational database.
- queryString: SQL query string.
- values: Array of values to be bound to SQL query string
- threshold: OPTIONAL. Threshold in milliseconds (ms). Default value is 3 seconds.
- TTL: OPTIONAL. Time to Live (TTL) in seconds (s). Default value is 30 mins.
- log: OPTIONAL. Turns console.logs on and off. Default value is false.
- instanceLatency: OPTIONAL. Switches from measuring total latency to measuring latency within the instance itself. Default value is false.
We recommend using the volatile-ttl eviction policy with this package to take advantage of the data's TTL, in the even that maximum memory is reached.
Session Storage
With an automatic cache backend architecture, a session store could be used to cache user session data, thereby improving the speed and efficiency of accessing session data since data is kept in a fast-access cache instead of slower primary storage, such as a relational database. The cache can store frequently accessed session data, which reduces the load on the primary database and improves response times for the user.
Machine Learning
Machine learning models can often be quite large and take time to load from a primary database. If an application needs to make frequent predictions in real time, it may be too slow to load the model from disk each time a prediction is needed. This NPM package can be used to cache critical components of the model in memory for fast access, significantly reducing the prediction latency and increasing the overall speed of the application.
Frequently Fetched Data or Costly Queries
Storing fetch requests' result sets in memory decreases the processing time for web applications that need to return data from complex queries. Furthermore, data that is frequently used need not be retrieved from a slow primary relational database, it can be kept in memory for quick retrieval as it is often requested by the client.
See the open issues for a full list of proposed features (and known issues).
Contributions are what make the open source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.
If you have a suggestion that would make this better, please fork the repo and create a pull request. You can also simply open an issue with the tag "enhancement". Don't forget to give the project a star! Thanks again!
- Fork the Project
- Create your Feature Branch (
git checkout -b feature/AmazingFeature
) - Commit your Changes (
git commit -m 'Add some AmazingFeature'
) - Push to the Branch (
git push origin feature/AmazingFeature
) - Open a Pull Request
Email - queryflow58@gmail.com
Twitter - @Query_Flow
LinkedIn - Team Page
QueryFlow Performance Visualizer and Analyzer: https://www.query-flow.com/
- Vivek Patel - GitHub - LinkedIn
- Niko Amescua - GitHub - LinkedIn
- Ryan Campbell - GitHub - LinkedIn
- Philip Brown - GitHub - LinkedIn
- George Greer - GitHub - LinkedIn
The Team wholeheartedly thanks Chris Suzukida for his mentorship and support throughout the development of this project.
Distributed under the MIT License. See LICENSE.txt
for more information.