hallucination-detection
There are 22 repositories under hallucination-detection topic.
uptrain-ai/uptrain
UpTrain is an open-source unified platform to evaluate and improve Generative AI applications. We provide grades for 20+ preconfigured checks (covering language, code, embedding use-cases), perform root cause analysis on failure cases and give insights on how to resolve them.
IAAR-Shanghai/UHGEval
[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
voidism/Lookback-Lens
Code for the EMNLP 2024 paper "Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps"
NishilBalar/Awesome-LVLM-Hallucination
up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resources
OpenKG-ORG/EasyDetect
An Easy-to-use Hallucination Detection Framework for LLMs.
deshwalmahesh/PHUDGE
Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.
open-compass/ANAH
[ACL 2024] ANAH & [NeurIPS 2024] ANAH-v2
patrick-tssn/VideoHallucer
VideoHallucer, The first comprehensive benchmark for hallucination detection in large video-language models (LVLMs)
zjunlp/EasyDetect
[ACL 2024] An Easy-to-use Hallucination Detection Framework for LLMs.
aimonlabs/aimon-python-sdk
This repo hosts the Python SDK and related examples for AIMon, which is a proprietary, state-of-the-art system for detecting LLM quality issues such as Hallucinations. It can be used during offline evals, continuous monitoring or inline detection. We offer various model quality metrics that are fast, reliable and cost-effective.
mbzuai-nlp/fire
Fact-checking with Iterative Retrieval and Verification
darveenvijayan/autoevaluator
Fully automated LLM evaluator
noanonkes/Hallucination-Detection-in-LLMs
Detecting Hallucinations in Large Language Model Generations using Graph Structures
BaluHarshavardan99/Hallucination-in-Chat-bots
Hallucination in Chat-bots: Faithful Benchmark for Information-Seeking Dialogue
ivan-kud/semeval-2024-shroom
Competition: SemEval-2024 Task-6 - SHROOM, a Shared-task on Hallucinations and Related Observable Overgeneration Mistakes
MuhammadTayyebHamid/Text-Classification-using-Logistic-Regression
Binary hallucination detection classifier using logistic regression
amarquaye/atlas
🔢Hallucination detector for Large Language Models.
amarquaye/atlas-api
API for the atlas project
amarquaye/atlas-chrome
Chrome extension for the ATLAS project.
jhaayush2004/RAG-Evaluation
Different approaches to evaluate RAG !!!
rafaelsandroni/antibodies
Antibodies for LLMs hallucinations (grouping LLM as a judge, NLI, reward models)