acewjh's Stars
HqWu-HITCS/Awesome-Chinese-LLM
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
opendilab/awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
PKU-Alignment/safe-rlhf
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
vectara/hallucination-leaderboard
Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents
snwfdhmp/awesome-gpt-prompt-engineering
A curated list of awesome resources, tools, and other shiny things for LLM prompt engineering.
Libr-AI/OpenFactVerification
Loki: Open-source solution designed to automate the process of verifying factuality
ydyjya/Awesome-LLM-Safety
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights into the safety implications, challenges, and advancements surrounding these powerful models.
ThuCCSLab/Awesome-LM-SSP
A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).
EdinburghNLP/awesome-hallucination-detection
List of papers on hallucination detection in LLMs.
google-deepmind/long-form-factuality
Benchmarking long-form factuality in large language models. Original code for our paper "Long-form factuality in large language models".
potsawee/selfcheckgpt
SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models
showlab/Awesome-MLLM-Hallucination
📖 A curated list of resources dedicated to hallucination of multimodal large language models (MLLM).
amazon-science/RefChecker
RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.
thu-coai/BPO
InternLM/OpenAOE
LLM Group Chat Framework: chat with multiple LLMs at the same time. 大模型群聊框架:同时与多个大语言模型聊天。
zjunlp/IEPile
[ACL 2024] IEPile: A Large-Scale Information Extraction Corpus
xieyuquanxx/awesome-Large-MultiModal-Hallucination
😎 up-to-date & curated list of awesome LMM hallucinations papers, methods & resources.
junyangwang0410/AMBER
An LLM-free Multi-dimensional Benchmark for Multi-modal Hallucination Evaluation
yinzhangyue/SelfAware
Do Large Language Models Know What They Don’t Know?
zjunlp/FactCHD
[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection
OpenKG-ORG/EasyDetect
An Easy-to-use Hallucination Detection Framework for LLMs.
javyduck/KnowHalu
terrierteam/pyterrier_doc2query
Ki-Seki/chat_prompt_templates
Collection of Basic Prompt Templates for Various Chat LLMs (Chat LLM 的基础提示模板集合)
AIFlames/Flames
Flames is a highly adversarial benchmark in Chinese for LLM's harmlessness evaluation developed by Shanghai AI Lab and Fudan NLP Group.
WhitzardIndex/WhitzardBench-2024A
复旦白泽大模型安全基准测试集(2024年夏季版)
webis-de/ACL-22
jiazhen-code/IntrinsicHallu
kinit-sk/disinformation-capabilities
Implementation of the paper "Disinformation Capabilities of Large Language Models"