fifiand1's Stars
ggerganov/llama.cpp
LLM inference in C/C++
aws-samples/bedrock-claude-chat
AWS-native chatbot using Bedrock + Claude (+Nova and Mistral)
microsoft/markitdown
Python tool for converting files and office documents to Markdown.
ollama/ollama-python
Ollama Python library
OpenBMB/MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
appium/appium-uiautomator2-server
Appium UiAutomator/UiObject2-based server for Android UI automation. This module is used by appium-uiautomator2-driver component
QwenLM/Qwen2.5-Coder
Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—language models
naginoa/LLMs_interview_notes
LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案
otterscan/otterscan
A blazingly fast, local, Ethereum block explorer built on top of Erigon
blockscout/blockscout
Blockchain explorer for Ethereum based network and a tool for inspecting and analyzing EVM based blockchains.
oobabooga/text-generation-webui
A Gradio web UI for Large Language Models with support for multiple inference backends.
unslothai/unsloth
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory
huggingface/trl
Train transformer language models with reinforcement learning.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
open-webui/open-webui
User-friendly AI Interface (Supports Ollama, OpenAI API, ...)
YiVal/YiVal
Your Automatic Prompt Engineering Assistant for GenAI Applications
langgenius/dify
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
chatchat-space/Langchain-Chatchat
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
NVIDIA/GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
xai-org/grok-1
Grok open release
ollama/ollama
Get up and running with Llama 3.3, Mistral, Gemma 2, and other large language models.
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
netease-youdao/QAnything
Question and Answer based on Anything.
QwenLM/qwen.cpp
C++ implementation of Qwen-LM
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
google-gemini/generative-ai-android
The official Android library for the Google Gemini API
jart/emacs-copilot
Large language model code completion for Emacs