Summary of Selective Preference Optimization Via Token-level Reward Function Estimation, by Kailai Yang et al.
Selective Preference Optimization via Token-Level Reward Function Estimationby Kailai Yang, Zhiwei Liu, Qianqian Xie, Jimin…
Selective Preference Optimization via Token-Level Reward Function Estimationby Kailai Yang, Zhiwei Liu, Qianqian Xie, Jimin…
MPruner: Optimizing Neural Network Size with CKA-Based Mutual Information Pruningby Seungbeom Hu, ChanJun Park, Andrew…
RuleAlign: Making Large Language Models Better Physicians with Diagnostic Rule Alignmentby Xiaohan Wang, Xiaoyan Yang,…
AlphaFolding: 4D Diffusion for Dynamic Protein Structure Prediction with Reference and Motion Guidanceby Kaihui Cheng,…
Weight Scope Alignment: A Frustratingly Easy Method for Model Mergingby Yichu Xu, Xin-Chun Li, Le…
Beyond Labels: Aligning Large Language Models with Human-like Reasoningby Muhammad Rafsan Kabir, Rafeed Mohammad Sultan,…
ProteinGPT: Multimodal LLM for Protein Property Prediction and Structure Understandingby Yijia Xiao, Edward Sun, Yiqiao…
Towards Aligned Data Removal via Twin Machine Unlearningby Yuyao Sun, Zhenxing Niu, Gang hua, Rong…
SEAL: Systematic Error Analysis for Value ALignmentby Manon Revel, Matteo Cargnelutti, Tyna Eloundou, Greg LeppertFirst…
Exploiting Fine-Grained Prototype Distribution for Boosting Unsupervised Class Incremental Learningby Jiaming Liu, Hongyuan Liu, Zhili…