Pinned Repositories
AI-ML-Research-Insights
This is a public repository of AI-ML Research Insights. Feel free to contribute.
model_server
A scalable inference server for models optimized with OpenVINO™
llama.cpp
LLM inference in C/C++
text-generation-inference
Large Language Model Text Generation Inference
airllm
AirLLM 70B inference with single 4GB GPU
ollama
Get up and running with Llama 3.3, Mistral, Gemma 2, and other large language models.
tinygrad
You like pytorch? You like micrograd? You love tinygrad! ❤️
udlbook
Understanding Deep Learning - Simon J.D. Prince
GPTCache
Semantic cache for LLMs. Fully integrated with LangChain and llama_index.
dhandhalyabhavik's Repositories
dhandhalyabhavik/AI-ML-Research-Insights
This is a public repository of AI-ML Research Insights. Feel free to contribute.
dhandhalyabhavik/model_server
A scalable inference server for models optimized with OpenVINO™