hallucination
There are 42 repositories under hallucination topic.
Awesome-LLM-Uncertainty-Reliability-Robustness
Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models
Woodpecker
✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models. The first work to correct hallucinations in MLLMs.
LRV-Instruction
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
RefChecker
RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.
HallusionBench
[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models
UHGEval
[ACL 2024] Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation
awesome-Large-MultiModal-Hallucination
😎 up-to-date & curated list of awesome LMM hallucinations papers, methods & resources.
TruthX
Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
FactCHD
[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection
LLaVA-Align
This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.
ICD
Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"
PHUDGE
Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.
LTI_Neural_Navigator
"Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases" by Jiarui Li and Ye Yuan and Zehua Zhang
NeurHal
Visual Correspondence Hallucination: Towards Geometric Reasoning (Under Review)
OLAPH
OLAPH: Improving Factuality in Biomedical Long-form Question Answering
KnowledgeCircuits
Knowledge Circuits in Pretrained Transformers
3D-GRAND
Official Implementation of 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs
KCA
Knowledge Verification to Nip Hallucination in the Bud
FaceAttr
CVPR2018 Face Super-resolution with supplementary Attributes
timechara
🧙🏻Code and benchmark for our Findings of ACL 2024 paper - "TimeChara: Evaluating Point-in-Time Character Hallucination of Role-Playing Large Language Models"
NLPCC2024_RegulatingLLM
[NLPCC 2024] Shared Task 10: Regulating Large Language Models
explainable-metrics
An explainable sentence similarity measurement
Skip-n
This repository contains the code of our paper 'Skip \n: A simple method to reduce hallucination in Large Vision-Language Models'.
EasyDetect
[ACL 2024] An Easy-to-use Hallucination Detection Framework for LLMs.
Awesome-LVLM-Hallucination
up-to-date and curated list of awesome state-of-the-art LVLMs hallucinations research work, papers & resources
CHALE
Controlled HALlucination-Evaluation (CHALE) Question-Answering Dataset
DD3412-chain-of-verification-reproduction
Re-implementation of the paper "Chain-of-Verification Reduces Hallucination in Large Language Models" for hallucination reduction. Developed as a final project of the Advanced Deep Learning course (DD3412) at KTH.
docker_openai_custom_weather_demo
openai function calling demo that gets customizable weather information
KCA
Knowledge Verification to Nip Hallucination in the Bud
DecodingStrategy
Unofficial implementations for optimized decoding strategies of large language models
dcpmi
[NAACL24] Official Implementation of Mitigating Hallucination in Abstractive Summarization with Domain-Conditional Mutual Information
SnackRiskAPI
The SnackRisk API logs potentially inaccurate outputs generated by GenAI bots and tools.
openai_assistant_code_interpreter
openai assistant using code interpreter
LLMVerify
Verify outputs generated by LLMs backed with real time data
lrec-coling-hallucination-tutorial
LREC-COLING 2024 Tutorial