Skip to content

Purshow/Awesome-LVLM-Hallucination

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

62 Commits
 
 

Repository files navigation

Awesome LVLM Hallucination

All the papers listed in this project come from my usual reading. If you have found some new and interesting papers, I would appreciate it if you let me know!!!

survey:

A Survey on Hallucination in Large Vision-Language Models

Hallucination of Multimodal Large Language Models: A Survey

Hallucination Benchmarks:

** arXiv Star

Mitigating(LVLM)

  1. LRV-Instruction: Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning, (Liu et al. ICLR2024)
    • Static Badge
    • [dataset] propose an instruction-tuning dataset that includes both positive and negative sample
    • GAIVE: evaluation approach which uses GPT-4
  2. LURE: Analyzing and Mitigating Object Hallucination in Large Vision-Language Models, (Zhou et al. ICLR2024)
    • Static Badge
    • [post-hoc revision] train a revision model to detect and correct hallucinated objects in the base model’s response.
  3. HallE-Switch: Rethinking and Controlling Object Existence Hallucinations in Large Vision-Language Models for Detailed Caption, (Zhai et al. 2023)
    • Static Badge
    • CCEval, a GPT-4 assisted evaluation method tailored for detailed captioning
  4. Woodpecker: Hallucination Correction for Multimodal Large Language Models, (Yin et al.)
    • Static Badge Static Badge
    • [revision] post-hoc correction
    • need other pre-trained visual models
  5. LLaVA-RLHF: Aligning Large Multimodal Models with Factually Augmented RLHF, (Sun et al.)
    • Static Badge
    • [RLHF-PPO] the first LMM trained with RLHF
    • propose benchmark: MMHal-Bench
  6. Volcano: Mitigating Multimodal Hallucination through Self-Feedback Guided Revision, (Lee et al.)
    • Static Badge
    • self-feedback, according to self-generate natural language feedback to self-revise response
  7. HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data, (Yu et al.)
    • Static Badge
  8. VCD: Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding, (Leng et al.)Highly recommended
    • Static Badge
    • constractive decoding
  9. HA-DPO: Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization
    • Static Badge Static Badge
  10. Mitigating Hallucination in Visual Language Models with Visual Supervision, (Chen et al.)
    • Static Badge Static Badge
  11. OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation, (Huang et al.)Highly recommended
    • Static Badge
    • Improve beam search
  12. FOHE: Mitigating Fine-Grained Hallucination by Fine-Tuning Large Vision-Language Models with Caption Rewrites, (Wang et al.)
    • Static Badge
    • use ChatGPT to post-hoc correction
  13. RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
    • Static Badge
    • [RLHF-DPO] 1.4K preference data, natural language feedback
  14. MOCHa: Multi-Objective Reinforcement Mitigating Caption Hallucinations, (Ben-Kish et al.)
    • Static Badge
    • [RLHF]
  15. HACL: Hallucination Augmented Contrastive Learning for Multimodal Large Language Model, (Jiang et al.)
    • Static Badge Static Badge
  16. Silkie: Preference Distillation for Large Visual Language Models, (Li et al.)
    • Static Badge
  17. HalluciDoctor:HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data:用“反常规”的数据干掉假相关性 - Static Badge
  18. MARINE: Mitigating Object Hallucination in LargeVision-Language Models via Classifier-Free Guidance
    • Static Badge
  19. CIEM :CIEM: Contrastive Instruction Evaluation Method for Better Instruction Tuning
    • Static Badge
  20. EFUF: EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models
    • Static Badge
  21. SEEING IS BELIEVING:MITIGATING HALLUCINATION IN LARGE VISION-LANGUAGE MODELS VIA CLIP-GUIDED DECODING
    • Static Badge
  22. Logical Closed Loop: Logical Closed Loop: Uncovering Object Hallucinations in Large Vision-Language Models
    • Static Badge
  23. Aligning Modalities in Vision Large Language Models via Preference Fine-tuning - Static Badge
  24. MOF:Eyes Wide Shut? Exploring the Visual Shortcomings of Multimodal LLMs 三幻神之作,强烈推荐!!!(这篇paper“养”活了好几篇paper.....)
    • Static Badge
  25. IBD: IBD: Alleviating Hallucinations in Large Vision-Language Models viaImage-Biased Decoding
    • Static Badge
  26. DualFocus:DualFocus: Integrating Macro and Micro Perspectives in Multi-modal Large Language Models
    • Static Badge
  27. HALC:HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding
    • Static Badge
  28. less is more:Less is More: Mitigating Multimodal Hallucination from an EOS Decision Perspective
    • Static Badge
  29. Number Hallucinations:Evaluating and Mitigating Number Hallucinations in Large Vision-Language Models: A Consistency Perspective
    • Static Badge 30-. Debiasing Large Visual Language Models Static Badge跟我想到的方法一模一样.....purshow最小丑的一集.....
  30. The First to Know: The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?
    • Static Badge
  31. Mitigating Dialogue Hallucination for Large Multi-modal Models via Adversarial Instruction Tuning
    • Static Badge
  32. Pensieve:Pensieve: Retrospect-then-Compare Mitigates Visual Hallucination
    • Static Badge
  33. Multi-Modal Hallucination Control by Visual Information Grounding
    • Static Badge
  34. What if...?:What if...?: Counterfactual Inception to Mitigate Hallucination Effects in Large Multimodal Models
    • Static Badge
  35. Exploiting Semantic Reconstruction to Mitigate Hallucinations in Vision-Language Models
    • Static Badge
  36. Mitigating Hallucinations in Large Vision-Language Models with Instruction Contrastive Decoding - Static Badge
  37. H2RSVLM:H2RSVLM: Towards Helpful and Honest Remote Sensing Large Vision Language Model
    • Static Badge
  38. FGAIF:FGAIF: Aligning Large Vision-Language Models with Fine-grained AI Feedback
    • Static Badge
  39. Joint Visual and Text Prompting for Improved Object-Centric Perception with Multimodal Large Language Models
    • Static Badge
  40. BRAVE:BRAVE: Broadening the visual encoding of vision-language models - Static Badge
  41. Learning to Localize Objects Improves Spatial Reasoning in Visual-LLMs
    • Static Badge
  42. LION:LION : Empowering Multimodal Large Language Model with Dual-Level Visual Knowledge
    • Static Badge
  43. Prescribing the Right Remedy: Prescribing the Right Remedy: Mitigating Hallucinations in Large Vision-Language Models via Targeted Instruction Tuning
    • Static Badge
  44. Fact:Fact :Teaching MLLMs with Faithful, Concise and Transferable Rationales
    • Static Badge
  45. Self-Supervised Visual Preference Alignment
    • Static Badge
  46. Exploring the Transferability of Visual Prompting for Multimodal Large Language Models
    • Static Badge
  47. VALOR-EVAL: VALOR-EVAL: Holistic Coverage and Faithfulness Evaluation of Large Vision-Language Models
    • Static Badge
  48. RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in LVLMs
    • Static Badge
  49. Mitigating Object Hallucination via Data Augmented Contrastive Tuning *
    • Static Badge
  50. Detecting Multimodal Situations with Insufficient Context and Abstaining from Baseless Predictions *
    • Static Badge
  51. Automated Multi-level Preference for MLLMs
    • Static Badge
  52. Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced Optimization - Static Badge
  53. VDGD VDGD: Mitigating LVLM Hallucinations in Cognitive Prompts by Bridging the Visual Perception Gap * - Static Badge
  54. Think Before You Act: A Two-Stage Framework for Mitigating Gender Bias Towards Vision-Language Tasks
    • Static Badge
  55. Don't Miss the Forest for the Trees Don't Miss the Forest for the Trees: Attentional Vision Calibration for Large Vision Language Models
    • Static Badge
  56. Calibrated Self-Rewarding Vision Language Models *
    • Static Badge
  57. VL-Uncertainty: Detecting Hallucination in Large Vision-Language Model via Uncertainty Estimation
    • Static Badge

他山之石:

survey(LLM's hallucination):

  • Song in the AI Ocean:A Survey on Hallucination in Large Language Models https://github.com/HillZhang1999/llm-hallucination-survey
  • A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions
  • A Comprehensive Survey of Hallucination Mitigation Techniques in LargeLanguage Models
  • A Survey of Hallucination in “Large” Foundation Models

Interesting topic:

解码层面:

ItI:Inference-Time Intervention:Eliciting Truthful Answers from a Language Model

CDS: Collaborative decoding of critical tokens for boosting factuality of large language models

Self-Consistent Decoding for More Factual Open Responses

Contrastive decoding: (vcd同款方法) I really think there are too many papers in this method....

  1. Contrastive Decoding: Open-ended Text Generation as Optimization Alleviating
  2. Alleviating Hallucinations of Large Language Models through Induced Halluctions
  3. Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
  4. SH2: Self-Highlighted Hesitation Helps You Decode More Truthfully
  5. DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models
  6. Discerning and Resolving Knowledge Conflicts through Adaptive Decoding with Contextual Information-Entropy Constraint
  7. ROSE Doesn’t Do That: Boosting the Safety of Instruction-Tuned Large Language Models with Reverse Prompt Contrastive Decoding
  8. Weak-to-Strong Jailbreaking on Large Language Models
  9. IBD: Alleviating Hallucinations in Large Vision-Language Models viaImage-Biased Decoding
  10. HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding
  11. ......

Acknowledgments

I am immensely grateful to two pivotal projects that have significantly influenced the development of my work: awesome-Large-MultiModal-Hallucination and Awesome-MLLM-Hallucination. The dedication and effort put forth by the contributors of these projects, particularly xieyuquanxx and the team at Show Lab, have provided an indispensable resource for researchers and developers alike. The awesome-Large-MultiModal-Hallucination project has offered a comprehensive guide and a curated list of resources that have been instrumental in shaping my understanding of Large MultiModal Hallucination. Similarly, the Awesome-MLLM-Hallucination repository has been a treasure trove of knowledge, showcasing cutting-edge techniques and methodologies in the realm of Machine Learning and Large Model Hallucination. By sharing their expertise and compiling these resources, they have not only advanced the field but also fostered a spirit of collaboration and open knowledge. I am deeply appreciative of their contributions and am inspired by their commitment to the community. Their work serves as a foundation upon which I have built and expanded, and for that, I extend my heartfelt thanks. This acknowledgment is a small gesture compared to the vast impact their work has had on mine. Thank you for setting a remarkable example for the open-source and scientific communities.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published