MarshtompCS's Stars
lyogavin/airllm
AirLLM 70B inference with single 4GB GPU
princeton-nlp/MeZO
[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
FlagAI-Open/FlagAI
FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.
IntelLabs/fastRAG
Efficient Retrieval Augmentation and Generation Framework
triton-lang/triton
Development repository for the Triton language and compiler
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
i-Eval/FairEval
openai/evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
composable-models/llm_multiagent_debate
ICML 2024: Improving Factuality and Reasoning in Language Models through Multiagent Debate
xlxwalex/FCGEC
The Corpus & Code for EMNLP 2022 paper "FCGEC: Fine-Grained Corpus for Chinese Grammatical Error Correction" | FCGEC中文语法纠错语料及STG模型
RUCAIBox/HaluEval
This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.
bigscience-workshop/promptsource
Toolkit for creating, sharing and using natural language prompts.
nickm980/smallville
Generative Agents for video games. Based on Generative Agents: Interactive Simulacra of Human Behavior
101dotxyz/GPTeam
GPTeam: An open-source multi-agent simulation
k2-fsa/icefall
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
hkust-nlp/ceval
Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]
FranxYao/GPT-Bargaining
Code for Arxiv 2023: Improving Language Model Negociation with Self-Play and In-Context Learning from AI Feedback
THU-KEG/EvaluationPapers4ChatGPT
Resource, Evaluation and Detection Papers for ChatGPT
terryyz/ice-score
[EACL 2024] ICE-Score: Instructing Large Language Models to Evaluate Code
OpenBMB/AgentVerse
🤖 AgentVerse 🪐 is designed to facilitate the deployment of multiple LLM-based agents in various applications, which primarily provides two frameworks: task-solving and simulation
CLUEbenchmark/SuperCLUE
SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
mymusise/ChatGLM-Tuning
基于ChatGLM-6B + LoRA的Fintune方案
Shikib/usr
Code for ACL 2020 paper: USR: An Unsupervised and Reference Free Evaluation Metric for Dialog Generation (https://arxiv.org/pdf/2005.00456)
maszhongming/UniEval
Repository for EMNLP 2022 Paper: Towards a Unified Multi-Dimensional Evaluator for Text Generation
f/awesome-chatgpt-prompts
This repo includes ChatGPT prompt curation to use ChatGPT better.
Yale-LILY/SummEval
Resources for the "SummEval: Re-evaluating Summarization Evaluation" paper
GasolSun36/Iter-CoT
[NAACL 2024] Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models
chatchat-space/Langchain-Chatchat
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and Llama) RAG and Agent app with langchain