Loading Now

Summary of A Survey Of Hallucination in Large Visual Language Models, by Wei Lan et al.


A Survey of Hallucination in Large Visual Language Models

by Wei Lan, Wenyi Chen, Qingfeng Chen, Shirui Pan, Huiyu Zhou, Yi Pan

First submitted to arxiv on: 20 Oct 2024

Categories

  • Main: Artificial Intelligence (cs.AI)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The Large Visual Language Models (LVLMs) integrate visual modality with Large Language Models (LLMs), enhancing user interaction and enriching user experience. LVLMs have demonstrated powerful information processing and generation capabilities, but hallucinations limit their potential and practical effectiveness. This survey reviews recent works on hallucination correction and mitigation, introducing the background of LVLMs and hallucinations, main causes of hallucination generation, and available evaluation benchmarks from judgmental and generative perspectives.
Low GrooveSquid.com (original content) Low Difficulty Summary
LVLMs combine visual and language processing to improve user interaction and experience. They can process and generate information well, but sometimes make things up (hallucinate). This makes it hard to trust them in real-life applications. This survey looks at how researchers are trying to fix this problem by correcting or mitigating hallucinations. It also explains why LVLMs work the way they do and what measures are used to test their reliability.

Keywords

» Artificial intelligence  » Hallucination