Summary of Q-groundcam: Quantifying Grounding in Vision Language Models Via Gradcam, by Navid Rajabi et al.
Q-GroundCAM: Quantifying Grounding in Vision Language Models via GradCAMby Navid Rajabi, Jana KoseckaFirst submitted to…
Q-GroundCAM: Quantifying Grounding in Vision Language Models via GradCAMby Navid Rajabi, Jana KoseckaFirst submitted to…
AAPL: Adding Attributes to Prompt Learning for Vision-Language Modelsby Gahyeon Kim, Sohee Kim, Seokju LeeFirst…
RetinaRegNet: A Zero-Shot Approach for Retinal Image Registrationby Vishal Balaji Sivaraman, Muhammad Imran, Qingyue Wei,…
MoDE: CLIP Data Experts via Clusteringby Jiawei Ma, Po-Yao Huang, Saining Xie, Shang-Wen Li, Luke…
SemEval-2024 Task 9: BRAINTEASER: A Novel Task Defying Common Senseby Yifan Jiang, Filip Ilievski, Kaixin…
Automated Multi-Language to English Machine Translation Using Generative Pre-Trained Transformersby Elijah Pelofske, Vincent Urias, Lorie…
FiLo: Zero-Shot Anomaly Detection by Fine-Grained Description and High-Quality Localizationby Zhaopeng Gu, Bingke Zhu, Guibo…
MARVEL: Multidimensional Abstraction and Reasoning through Visual Evaluation and Learningby Yifan Jiang, Jiarui Zhang, Kexuan…
UnibucLLM: Harnessing LLMs for Automated Prediction of Item Difficulty and Response Time for Multiple-Choice Questionsby…
Data Alignment for Zero-Shot Concept Generation in Dermatology AIby Soham Gadgil, Mahtab BigverdiFirst submitted to…