Summary of How Well Do Multi-modal Llms Interpret Ct Scans? An Auto-evaluation Framework For Analyses, by Qingqing Zhu et al.
How Well Do Multi-modal LLMs Interpret CT Scans? An Auto-Evaluation Framework for Analysesby Qingqing Zhu,…
How Well Do Multi-modal LLMs Interpret CT Scans? An Auto-Evaluation Framework for Analysesby Qingqing Zhu,…
GPT as Psychologist? Preliminary Evaluations for GPT-4V on Visual Affective Computingby Hao Lu, Xuesong Niu,…
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generationby Zihao Wang, Anji Liu, Haowei…
An In-depth Evaluation of GPT-4 in Sentence Simplification with Error-based Human Assessmentby Xuanxin Wu, Yuki…
Feedback-Generation for Programming Exercises With GPT-4by Imen Azaiz, Natalie Kiesler, Sven StrickrothFirst submitted to arxiv…
Guiding Enumerative Program Synthesis with Large Language Modelsby Yixuan Li, Julian Parsert, Elizabeth PolgreenFirst submitted…
Can Large Language Models do Analytical Reasoning?by Yebowen Hu, Kaiqiang Song, Sangwoo Cho, Xiaoyang Wang,…
Emotional Manipulation Through Prompt Engineering Amplifies Disinformation Generation in AI Large Language Modelsby Rasita Vinay,…
Assessing the Aesthetic Evaluation Capabilities of GPT-4 with Vision: Insights from Group and Individual Assessmentsby…
Multimodal Large Language Models to Support Real-World Fact-Checkingby Jiahui Geng, Yova Kementchedjhieva, Preslav Nakov, Iryna…