hallucination

There are 42 repositories under hallucination topic.

  • DenseLidarNet

    Language:Jupyter Notebook34
  • Awesome-LLM-Uncertainty-Reliability-Robustness

    Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models

  • Woodpecker

    ✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models. The first work to correct hallucinations in MLLMs.

    Language:Python572
  • LRV-Instruction

    [ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning

    Language:Python233
  • RefChecker

    RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.

    Language:Python225
  • HallusionBench

    [CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models

    Language:Python202
  • UHGEval

    [ACL 2024] Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation

    Language:Python166
  • awesome-Large-MultiModal-Hallucination

    😎 up-to-date & curated list of awesome LMM hallucinations papers, methods & resources.

  • TruthX

    Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"

    Language:Python78
  • FactCHD

    [IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection

    Language:Python72
  • LLaVA-Align

    This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.

    Language:Python65
  • ICD

    Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"

    Language:Python55
  • PHUDGE

    Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.

    Language:Jupyter Notebook44
  • LTI_Neural_Navigator

    "Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases" by Jiarui Li and Ye Yuan and Zehua Zhang

    Language:HTML31
  • NeurHal

    Visual Correspondence Hallucination: Towards Geometric Reasoning (Under Review)

  • OLAPH

    OLAPH: Improving Factuality in Biomedical Long-form Question Answering

    Language:Python28
  • KnowledgeCircuits

    Knowledge Circuits in Pretrained Transformers

    Language:Python28
  • 3D-GRAND

    Official Implementation of 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs

  • KCA

    Knowledge Verification to Nip Hallucination in the Bud

    Language:Python18
  • FaceAttr

    CVPR2018 Face Super-resolution with supplementary Attributes

    Language:Lua18
  • timechara

    🧙🏻Code and benchmark for our Findings of ACL 2024 paper - "TimeChara: Evaluating Point-in-Time Character Hallucination of Role-Playing Large Language Models"

    Language:Python13
  • NLPCC2024_RegulatingLLM

    [NLPCC 2024] Shared Task 10: Regulating Large Language Models

  • explainable-metrics

    An explainable sentence similarity measurement

    Language:Jupyter Notebook11
  • Skip-n

    This repository contains the code of our paper 'Skip \n: A simple method to reduce hallucination in Large Vision-Language Models'.

    Language:Python10
  • EasyDetect

    [ACL 2024] An Easy-to-use Hallucination Detection Framework for LLMs.

    Language:Python8
  • Awesome-LVLM-Hallucination

    up-to-date and curated list of awesome state-of-the-art LVLMs hallucinations research work, papers & resources

  • CHALE

    Controlled HALlucination-Evaluation (CHALE) Question-Answering Dataset

    Language:Python3
  • DD3412-chain-of-verification-reproduction

    Re-implementation of the paper "Chain-of-Verification Reduces Hallucination in Large Language Models" for hallucination reduction. Developed as a final project of the Advanced Deep Learning course (DD3412) at KTH.

    Language:Python3
  • docker_openai_custom_weather_demo

    docker_openai_custom_weather_demo

    openai function calling demo that gets customizable weather information

    Language:Python2
  • KCA

    Knowledge Verification to Nip Hallucination in the Bud

    Language:Python1
  • DecodingStrategy

    Unofficial implementations for optimized decoding strategies of large language models

    Language:Jupyter Notebook1
  • dcpmi

    [NAACL24] Official Implementation of Mitigating Hallucination in Abstractive Summarization with Domain-Conditional Mutual Information

    Language:Python1
  • SnackRiskAPI

    The SnackRisk API logs potentially inaccurate outputs generated by GenAI bots and tools.

    Language:Go
  • openai_assistant_code_interpreter

    openai assistant using code interpreter

    Language:Python
  • LLMVerify

    Verify outputs generated by LLMs backed with real time data

    Language:Python
  • lrec-coling-hallucination-tutorial

    LREC-COLING 2024 Tutorial

    Language:JavaScript