SylvanLiu's Stars
ggerganov/llama.cpp
LLM inference in C/C++
lutzroeder/netron
Visualizer for neural network, deep learning and machine learning models
Vision-CAIR/MiniGPT-4
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
OpenBMB/MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
artidoro/qlora
QLoRA: Efficient Finetuning of Quantized LLMs
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
onnx/models
A collection of pre-trained, state-of-the-art models in the ONNX format
SJTU-IPADS/PowerInfer
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
yangjianxin1/Firefly
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
AutoGPTQ/AutoGPTQ
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
NExT-GPT/NExT-GPT
Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model
LLaVA-VL/LLaVA-NeXT
straight-tamago/misakaX
iOS /iPadOS 16.0 - 18.0 / 18.1 beta 4, An ultimate customization tool, uilitizing the bug that makes TrollRestore possible.
PhoebusSi/Alpaca-CoT
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
Kyle-Ye/XcodeLLMEligible
haoheliu/AudioLDM
AudioLDM: Generate speech, sound effects, music and beyond, with text.
HarderThenHarder/transformers_tasks
⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SFT etc.
casper-hansen/AutoAWQ
AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:
GAIR-NLP/O1-Journey
O1 Replication Journey: A Strategic Progress Report – Part I
digital-standard/ThreeDPoseUnityBarracuda
Unity sample of 3D pose estimation using Barracuda
Toyhom/Chinese-medical-dialogue-data
Chinese medical dialogue data 中文医疗对话数据集
NiuTrans/Classical-Modern
非常全的文言文(古文)-现代文平行语料
kakaobrain/coyo-dataset
COYO-700M: Large-scale Image-Text Pair Dataset
SimpleBerry/LLaMA-O1
Large Reasoning Models
google-research/long-range-arena
Long Range Arena for Benchmarking Efficient Transformers
shuxueslpi/chatGLM-6B-QLoRA
使用peft库,对chatGLM-6B/chatGLM2-6B实现4bit的QLoRA高效微调,并做lora model和base model的merge及4bit的量化(quantize)。
YuxiXie/MCTS-DPO
This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.