Summary of Eyeclip: a Visual-language Foundation Model For Multi-modal Ophthalmic Image Analysis, by Danli Shi et al.
EyeCLIP: A visual-language foundation model for multi-modal ophthalmic image analysisby Danli Shi, Weiyi Zhang, Jiancheng…
EyeCLIP: A visual-language foundation model for multi-modal ophthalmic image analysisby Danli Shi, Weiyi Zhang, Jiancheng…
Quantifying and Enabling the Interpretability of CLIP-like Modelsby Avinash Madasu, Yossi Gandelsman, Vasudev Lal, Phillip…
LLaMA-Omni: Seamless Speech Interaction with Large Language Modelsby Qingkai Fang, Shoutao Guo, Yan Zhou, Zhengrui…
Semifactual Explanations for Reinforcement Learningby Jasmina Gajcin, Jovan Jeromela, Ivana DusparicFirst submitted to arxiv on:…
AD-Net: Attention-based dilated convolutional residual network with guided decoder for robust skin lesion segmentationby Asim…
Proto-OOD: Enhancing OOD Object Detection with Prototype Feature Similarityby Junkun Chen, Jilin Mei, Liang Chen,…
Elsevier Arena: Human Evaluation of Chemistry/Biology/Health Foundational Large Language Modelsby Camilo Thorne, Christian Druckenbrodt, Kinga…
Harmonic Reasoning in Large Language Modelsby Anna KruspeFirst submitted to arxiv on: 9 Sep 2024CategoriesMain:…
Seeing Through the Mask: Rethinking Adversarial Examples for CAPTCHAsby Yahya Jabary, Andreas Plesner, Turlan Kuzhagaliyev,…
HMAFlow: Learning More Accurate Optical Flow via Hierarchical Motion Field Alignmentby Dianbo Ma, Kousuke Imamura,…