Summary of Phd: a Chatgpt-prompted Visual Hallucination Evaluation Dataset, by Jiazhen Liu et al.
PhD: A ChatGPT-Prompted Visual hallucination Evaluation Datasetby Jiazhen Liu, Yuhan Fu, Ruobing Xie, Runquan Xie,…
PhD: A ChatGPT-Prompted Visual hallucination Evaluation Datasetby Jiazhen Liu, Yuhan Fu, Ruobing Xie, Runquan Xie,…
Unsupervised Real-Time Hallucination Detection based on the Internal States of Large Language Modelsby Weihang Su,…
On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in Summarizationby Lorenzo…
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generationby Zihao Wang, Anji Liu, Haowei…
ChatASU: Evoking LLM’s Reflexion to Truly Understand Aspect Sentiment in Dialoguesby Yiding Liu, Jingjing Wang,…
German also Hallucinates! Inconsistency Detection in News Summaries with the Absinth Datasetby Laura Mascarell, Ribin…
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language Modelsby Kedi Chen, Qin Chen, Jie…
Whispers that Shake Foundations: Analyzing and Mitigating False Premise Hallucinations in Large Language Modelsby Hongbang…
Editing Factual Knowledge and Explanatory Ability of Medical Large Language Modelsby Derong Xu, Ziheng Zhang,…
Re-Ex: Revising after Explanation Reduces the Factual Errors in LLM Responsesby Juyeon Kim, Jeongeun Lee,…