This repository will be updated to include all the contemporary papers related to hallucination in foundation models. We broadly categorize the papers into four major categories as follows.
- SELFCHECKGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models
- Beyond Factuality: A Comprehensive Evaluation of Large Language Models as Knowledge Generators
- HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language Models
- Self-contradictory Hallucinations of Large Language Models: Evaluation, Detection and Mitigation
- PURR: Efficiently Editing Language Model Hallucinations by Denoising Language Model Corruptions
- Mitigating Language Model Hallucination with Interactive Question-Knowledge Alignment
- How Language Model Hallucinations Can Snowball
- Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback
- The Internal State of an LLM Knows When its Lying
- Chain of Knowledge: A Framework for Grounding Large Language Models with Structured Knowledge Bases
- HALO: Estimation and Reduction of Hallucinations in Open-Source Weak Large Language Models
- A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation
- Dehallucinating Large Language Models Using Formal Methods Guided Iterative Prompting
- Sources of Hallucination by Large Language Models on Inference Tasks
- Citation: A Key to Building Responsible and Accountable Large Language Models
- Zero-resource hallucination prevention for large language models
- RARR: Researching and Revising What Language Models Say, Using Language Models
- Evaluating Object Hallucination in Large Vision-Language Models
- Detecting and Preventing Hallucinations in Large Vision Language Models
- Plausible May Not Be Faithful: Probing Object Hallucination in Vision-Language Pre-training
- Hallucination Improves the Performance of Unsupervised Visual Representation Learning