Enhancing Reliability in Multimodal Models

The recent advancements in Large Vision Language Models (LVLMs) have significantly enhanced multimodal understanding and interaction. However, these models often suffer from hallucinations, where the generated outputs do not accurately reflect the input data, particularly in visual and audio-visual contexts. The field is currently focused on developing techniques to mitigate these hallucinations, with a particular emphasis on improving the alignment between visual inputs and textual outputs. Novel methods such as latent space steering and concentric causal attention are being explored to stabilize vision features and reduce the sensitivity of text decoders to visual inputs. Additionally, the introduction of specialized benchmarks like AVHBench is crucial for evaluating and improving the perception and comprehension capabilities of audio-visual LLMs. These developments highlight a shift towards more robust and reliable multimodal models, aiming to enhance their practical applicability across various domains.

Noteworthy papers include one that introduces Visual and Textual Intervention (VTI) to steer latent space representations, effectively reducing hallucinations, and another that proposes Concentric Causal Attention (CCA) to mitigate object hallucination by improving positional alignment between visual and instruction tokens.

Sources

A Survey of Hallucination in Large Visual Language Models

Reducing Hallucinations in Vision-Language Models via Latent Space Steering

Mitigating Object Hallucination via Concentric Causal Attention

AVHBench: A Cross-Modal Hallucination Benchmark for Audio-Visual Large Language Models

Built with on top of