ObserverWho's Stars
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
ollama/ollama
Get up and running with Llama 3.1, Mistral, Gemma 2, and other large language models.
exacity/deeplearningbook-chinese
Deep Learning Book Chinese Translation
Chanzhaoyu/chatgpt-web
用 Express 和 Vue3 搭建的 ChatGPT 演示网页
chatchat-space/Langchain-Chatchat
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain
milvus-io/milvus
A cloud-native vector database, storage for next generation AI applications
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
hpcaitech/Open-Sora
Open-Sora: Democratizing Efficient Video Production for All
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
joaomdmoura/crewAI
Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
vernesong/OpenClash
A Clash Client For OpenWrt
eosphoros-ai/DB-GPT
AI Native Data App Development framework with AWEL(Agentic Workflow Expression Language) and Agents
netease-youdao/QAnything
Question and Answer based on Anything.
juicedata/juicefs
JuiceFS is a distributed POSIX file system built on top of Redis and S3.
kedacore/keda
KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
activeloopai/deeplake
Database for AI. Store Vectors, Images, Texts, Videos, etc. Use with LLMs/LangChain. Store, query, version, & visualize any AI data. Stream data in real-time to PyTorch/TensorFlow. https://activeloop.ai
triton-inference-server/server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
datawhalechina/self-llm
《开源大模型食用指南》基于Linux环境快速部署开源大模型,更适合**宝宝的部署教程
kyrolabs/awesome-langchain
😎 Awesome list of tools and projects with the awesome LangChain framework
bentoml/BentoML
The easiest way to serve AI apps and models - Build reliable Inference APIs, LLM apps, Multi-model chains, RAG service, and much more!
NVIDIA/cuda-samples
Samples for CUDA Developers which demonstrates features in CUDA Toolkit
uber/kraken
P2P Docker registry capable of distributing TBs of data in seconds
k8sgpt-ai/k8sgpt
Giving Kubernetes Superpowers to everyone
datawhalechina/llm-universe
本项目是一个面向小白开发者的大模型应用开发教程,在线阅读地址:https://datawhalechina.github.io/llm-universe/
hwchase17/langchain-hub
jeinlee1991/chinese-llm-benchmark
中文大模型能力评测榜单:目前已囊括115个大模型,覆盖chatgpt、gpt4o、百度文心一言、阿里通义千问、讯飞星火、商汤senseChat、minimax等商用模型, 以及百川、qwen2、glm4、yi、书生internLM2、llama3等开源大模型,多维度能力评测。不仅提供能力评分排行榜,也提供所有模型的原始输出结果!
dragonflyoss/Dragonfly2
Dragonfly is an open source P2P-based file distribution and image acceleration system. It is hosted by the Cloud Native Computing Foundation (CNCF) as an Incubating Level Project.
ray-project/kuberay
A toolkit to run Ray applications on Kubernetes
huataihuang/cloud-atlas
云计算指南