Summary of Cascade-aware Training Of Language Models, by Congchao Wang et al.
Cascade-Aware Training of Language Modelsby Congchao Wang, Sean Augenstein, Keith Rush, Wittawat Jitkrittum, Harikrishna Narasimhan,…
Cascade-Aware Training of Language Modelsby Congchao Wang, Sean Augenstein, Keith Rush, Wittawat Jitkrittum, Harikrishna Narasimhan,…
Exploratory Preference Optimization: Harnessing Implicit Q*-Approximation for Sample-Efficient RLHFby Tengyang Xie, Dylan J. Foster, Akshay…
What Makes CLIP More Robust to Long-Tailed Pre-Training Data? A Controlled Study for Transferable Insightsby…
PostDoc: Generating Poster from a Long Multimodal Document Using Deep Submodular Optimizationby Vijay Jaisankar, Sambaran…
CLIPLoss and Norm-Based Data Selection Methods for Multimodal Contrastive Learningby Yiping Wang, Yifang Chen, Wendan…
Robust Preference Optimization through Reward Model Distillationby Adam Fisch, Jacob Eisenstein, Vicky Zayats, Alekh Agarwal,…
MAP-Neo: Highly Capable and Transparent Bilingual Large Language Model Seriesby Ge Zhang, Scott Qu, Jiaheng…
Robust Optimization in Protein Fitness Landscapes Using Reinforcement Learning in Latent Spaceby Minji Lee, Luiz…
Knowledge Circuits in Pretrained Transformersby Yunzhi Yao, Ningyu Zhang, Zekun Xi, Mengru Wang, Ziwen Xu,…
CLAIM Your Data: Enhancing Imputation Accuracy with Contextual Large Language Modelsby Ahatsham Hayat, Mohammad Rashedul…