Summary of Reformatted Alignment, by Run-ze Fan et al.
Reformatted Alignmentby Run-Ze Fan, Xuefeng Li, Haoyang Zou, Junlong Li, Shwai He, Ethan Chern, Jiewen…
Reformatted Alignmentby Run-Ze Fan, Xuefeng Li, Haoyang Zou, Junlong Li, Shwai He, Ethan Chern, Jiewen…
AnyGPT: Unified Multimodal LLM with Discrete Sequence Modelingby Jun Zhan, Junqi Dai, Jiasheng Ye, Yunhua…
Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmarkby Yihua Zhang, Pingzhi Li, Junyuan Hong,…
Aligning Modalities in Vision Large Language Models via Preference Fine-tuningby Yiyang Zhou, Chenhang Cui, Rafael…
Aligning Large Language Models by On-Policy Self-Judgmentby Sangkyu Lee, Sungdong Kim, Ashkan Yousefpour, Minjoon Seo,…
Multi-modal Preference Alignment Remedies Degradation of Visual Instruction Tuning on Language Modelsby Shengzhi Li, Rongyu…
Active Preference Optimization for Sample Efficient RLHFby Nirjhar Das, Souradip Chakraborty, Aldo Pacchiano, Sayak Ray…
Efficient Generative Modeling via Penalized Optimal Transport Networkby Wenhui Sophia Lu, Chenyang Zhong, Wing Hung…
Pretext Training Algorithms for Event Sequence Databy Yimu Wang, He Zhao, Ruizhi Deng, Frederick Tung,…
Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustmentby Rui Yang, Xiaoman Pan, Feng…