Loading Now

Summary of Xgen-mm (blip-3): a Family Of Open Large Multimodal Models, by Le Xue et al.


xGen-MM (BLIP-3): A Family of Open Large Multimodal Models

by Le Xue, Manli Shu, Anas Awadalla, Jun Wang, An Yan, Senthil Purushwalkam, Honglu Zhou, Viraj Prabhu, Yutong Dai, Michael S Ryoo, Shrikant Kendre, Jieyu Zhang, Can Qin, Shu Zhang, Chia-Chih Chen, Ning Yu, Juntao Tan, Tulika Manoj Awalgaonkar, Shelby Heinecke, Huan Wang, Yejin Choi, Ludwig Schmidt, Zeyuan Chen, Silvio Savarese, Juan Carlos Niebles, Caiming Xiong, Ran Xu

First submitted to arxiv on: 16 Aug 2024

Categories

  • Main: Computer Vision and Pattern Recognition (cs.CV)
  • Secondary: Artificial Intelligence (cs.AI); Computation and Language (cs.CL)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This report introduces xGen-MM (also known as BLIP-3), a framework for developing Large Multimodal Models (LMMs). The framework comprises meticulously curated datasets, a training recipe, model architectures, and a resulting suite of LMMs. xGen-MM expands the Salesforce xGen initiative on foundation AI models. Our pre-trained base model exhibits strong in-context learning capabilities, while the instruction-tuned model demonstrates competitive performance among open-source LMMs with similar model sizes. Additionally, we introduce a safety-tuned model with DPO to mitigate harmful behaviors such as hallucinations and improve safety. We also release our models, curated large-scale datasets, and fine-tuning codebase to facilitate further advancements in LMM research.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper creates a new way to make Large Multimodal Models (LMMs) using xGen-MM. This framework has many parts, including special datasets, a recipe for training, and different model designs. The goal is to help make better AI models that can learn from lots of information. The pre-trained model does well at learning in new situations, while the instruction-tuned model performs similarly to other open-source LMMs with similar sizes. They also made a safety-tuned model to prevent bad behaviors and improve safety.

Keywords

» Artificial intelligence  » Fine tuning