Loading Now

Summary of Freeze the Backbones: a Parameter-efficient Contrastive Approach to Robust Medical Vision-language Pre-training, by Jiuming Qin et al.


Freeze the backbones: A Parameter-Efficient Contrastive Approach to Robust Medical Vision-Language Pre-training

by Jiuming Qin, Che Liu, Sibo Cheng, Yike Guo, Rossella Arcucci

First submitted to arxiv on: 2 Jan 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Artificial Intelligence (cs.AI); Machine Learning (cs.LG)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed Adaptor framework addresses the limitations of existing Vision-Language Self-Supervised Learning (VL-SSL) approaches by preserving prior knowledge in pre-trained encoders while reducing computation requirements. By freezing pre-trained image and text encoders and employing a lightweight Adaptor module for cross-modal learning, the framework achieves competitive performance on medical image classification and segmentation tasks across three datasets, with trainable parameters reduced by over 90% compared to current pre-training approaches.
Low GrooveSquid.com (original content) Low Difficulty Summary
The Adaptor framework is a new way of learning from radiographic images and textual reports. It helps doctors diagnose patients better. The idea is to keep using big models that were already trained on lots of data, and then add a small extra part that makes the model learn from medical images. This makes it faster and more efficient. The results show that this new approach works well and can even be used with very little training data.

Keywords

* Artificial intelligence  * Image classification  * Self supervised