semantic-cache
There are 13 repositories under semantic-cache topic.
codefuse-ai/ModelCache
A LLM semantic caching system aiming to enhance user experience by reducing response time via cached query-result pairs.
redis/redis-vl-python
Redis Vector Library (RedisVL) -- the AI-native Python client for Redis.
sensoris/semcache
Semantic caching layer for your LLM applications. Reuse responses and reduce token usage.
vcache-project/vCache
Reliable and Efficient Semantic Prompt Caching with vCache
Harras3/Enterprise-Grade-RAG
This is a RAG based chatbot in which semantic cache and guardrails have been incorporated.
aws-samples/Reducing-Hallucinations-in-LLM-Agents-with-a-Verified-Semantic-Cache
This repository contains sample code demonstrating how to implement a verified semantic cache using Amazon Bedrock Knowledge Bases to prevent hallucinations in Large Language Model (LLM) responses while improving latency and reducing costs.
jonathanscholtes/LLM-Performance-with-Azure-Cosmos-DB-Semantic-Cache
Enhance LLM retrieval performance with Azure Cosmos DB Semantic Cache. Learn how to integrate and optimize caching strategies in real-world web applications.
mar1boroman/redis-movies-gen-ai
Redis Vector Similarity Search, Semantic Caching, Recommendation Systems and RAG
benitomartin/semantic-caching-qdrant-splade
Optimized RAG Retrieval with Indexing, Quantization, Hybrid Search and Caching
mar1boroman/ask-redis-blogs
A ChatBot using Redis Vector Similarity Search, which can recommend blogs based on user prompt
mar1boroman/text-sql-chatbot
Redis Database offers unique capability to keep your data fresh while serving through LLM chatbot
semioz/simcache
Semantic cache for your LLM apps in Go!
paswell-chiks/Optimizing-RAG-with-Hybrid-Search
🔍 Optimize RAG systems by exploring Lexical, Semantic, and Hybrid Search methods for better context retrieval and improved LLM responses.