BeyonderXX
PhD student at Fudan NLP. Actively looking for postdoc positions!!
Fudan UniversityShanghai, China
BeyonderXX's Stars
ollama/ollama
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
ggerganov/llama.cpp
LLM inference in C/C++
ray-project/ray
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—foundation models
openai/tiktoken
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
THUDM/GLM-4
GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
open-compass/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
llm-attacks/llm-attacks
Universal and Transferable Attacks on Aligned Language Models
opendilab/awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
VainF/Torch-Pruning
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
OpenLLMAI/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
SkyworkAI/Skywork
Skywork series models are pre-trained on 3.2TB of high-quality multilingual (mainly Chinese and English) and code data. We have open-sourced the model, training data, evaluation data, evaluation methods, etc. 天工系列模型在3.2TB高质量多语言和代码数据上进行预训练。我们开源了模型参数,训练数据,评估数据,评估方法。
master131/iFakeLocation
Simulate locations on iOS devices on Windows, Mac and Ubuntu.
thu-coai/Safety-Prompts
Chinese safety prompts for evaluating and improving the safety of LLMs. 中文安全prompts,用于评估和提升大模型的安全性。
horseee/LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
stone-zeng/fduthesis
LaTeX thesis template for Fudan University
IST-DASLab/sparsegpt
Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".
GaryYufei/AlignLLMHumanSurvey
Aligning Large Language Models with Human: A Survey
BradyFU/Woodpecker
✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models. The first work to correct hallucinations in MLLMs.
EasyJailbreak/EasyJailbreak
An easy-to-use Python framework to generate adversarial jailbreak prompts.
llava-rlhf/LLaVA-RLHF
Aligning LMMs with Factually Augmented RLHF
zhijing-jin/NLP4SocialGood_Papers
A reading list of up-to-date papers on NLP for Social Good.
HuangLK/transpeeder
train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism
SkyworkAI/Skywork-MoE
Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models
BeyonderXX/TRACE
TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models
SparkJiao/llama-pipeline-parallel
A prototype repo for hybrid training of pipeline parallel and distributed data parallel with comments on core code snippets. Feel free to copy code and launch discussions about the problems you have encoured.
shreyansh26/Red-Teaming-Language-Models-with-Language-Models
A re-implementation of the "Red Teaming Language Models with Language Models" paper by Perez et al., 2022
huizhang-L/CodeChameleon
InvokerStark/OverKill