Summary of Lire: Listwise Reward Enhancement For Preference Alignment, by Mingye Zhu et al.
LIRE: listwise reward enhancement for preference alignmentby Mingye Zhu, Yi Liu, Lei Zhang, Junbo Guo,…
LIRE: listwise reward enhancement for preference alignmentby Mingye Zhu, Yi Liu, Lei Zhang, Junbo Guo,…
Pragmatic auditing: a pilot-driven approach for auditing Machine Learning systemsby Djalel Benbouzid, Christiane Plociennik, Laura…
Energy Rank Alignment: Using Preference Optimization to Search Chemical Space at Scaleby Shriram Chennakesavalu, Frank…
Exploring and Exploiting the Asymmetric Valley of Deep Neural Networksby Xin-Chun Li, Jin-Lin Tang, Bo…
SPO: Multi-Dimensional Preference Sequential Alignment With Implicit Reward Modelingby Xingzhou Lou, Junge Zhang, Jian Xie,…
Hummer: Towards Limited Competitive Preference Datasetby Li Jiang, Yusen Wu, Junwu Xiong, Jingqing Ruan, Yichuan…
Reproducibility Study of CDUL: CLIP-Driven Unsupervised Learning for Multi-Label Image Classificationby Manan Shah, Yash BhalgatFirst…
OpenRLHF: An Easy-to-use, Scalable and High-performance RLHF Frameworkby Jian Hu, Xibin Wu, Zilin Zhu, Xianyu,…
Conformal Alignment: Knowing When to Trust Foundation Models with Guaranteesby Yu Gui, Ying Jin, Zhimei…
Spectral Editing of Activations for Large Language Model Alignmentby Yifu Qiu, Zheng Zhao, Yftah Ziser,…