tsingcoo's Stars
chroma-core/chroma
the AI-native open-source embedding database
ZNLP/BigTranslate
BigTranslate: Augmenting Large Language Models with Multilingual Translation Capability over 100 Languages
lemon0830/TIM
code for Teaching LM to Translate with Comparison
wxjiao/ParroT
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
deep-spin/hallucinations-in-nmt
hongbinye/Cognitive-Mirage-Hallucinations-in-LLMs
Repository for the paper "Cognitive Mirage: A Review of Hallucinations in Large Language Models"
lancopku/label-words-are-anchors
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
FreedomIntelligence/GrammarGPT
The code and data for GrammarGPT.
HillZhang1999/llm-hallucination-survey
Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large Language Models"
HillZhang1999/MuCGEC
MuCGEC中文纠错数据集及文本纠错SOTA模型开源;Code & Data for our NAACL 2022 Paper "MuCGEC: a Multi-Reference Multi-Source Evaluation Dataset for Chinese Grammatical Error Correction"
DefTruth/Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
kimiyoung/transformer-xl
bojone/bytepiece
更纯粹、更高压缩率的Tokenizer
intro-llm/intro-llm.github.io
website
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
baichuan-inc/Baichuan2
A series of large language models developed by Baichuan Intelligent Technology
IST-DASLab/gptq
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
QingruZhang/AdaLoRA
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
microsoft/gpt-MT
xverse-ai/XVERSE-13B
XVERSE-13B: A multilingual large language model developed by XVERSE Technology Inc.
Unbabel/COMET
A Neural Framework for MT Evaluation
QwenLM/Qwen
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
fkodom/grouped-query-attention-pytorch
(Unofficial) PyTorch implementation of grouped-query attention (GQA) from "GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints" (https://arxiv.org/pdf/2305.13245.pdf)
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
explosion/spaCy
💫 Industrial-strength Natural Language Processing (NLP) in Python
InternLM/lmdeploy
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
DAMO-NLP-MT/PolyLM
ggerganov/llama.cpp
LLM inference in C/C++