Summary of Robust Preference Optimization Through Reward Model Distillation, by Adam Fisch et al.
Robust Preference Optimization through Reward Model Distillationby Adam Fisch, Jacob Eisenstein, Vicky Zayats, Alekh Agarwal,…
Robust Preference Optimization through Reward Model Distillationby Adam Fisch, Jacob Eisenstein, Vicky Zayats, Alekh Agarwal,…
Provable Contrastive Continual Learningby Yichen Wen, Zhiquan Tan, Kaipeng Zheng, Chuanlong Xie, Weiran HuangFirst submitted…
FedHPL: Efficient Heterogeneous Federated Learning with Prompt Tuning and Logit Distillationby Yuting Ma, Lechao Cheng,…
EM Distillation for One-step Diffusion Modelsby Sirui Xie, Zhisheng Xiao, Diederik P Kingma, Tingbo Hou,…
Enhancing Consistency-Based Image Generation via Adversarialy-Trained Classification and Energy-Based Discriminationby Shelly Golan, Roy Ganz, Michael…
Score Distillation via Reparametrized DDIMby Artem Lukoianov, Haitz Sáez de Ocáriz Borde, Kristjan Greenewald, Vitor…
Interpretable and Editable Programmatic Tree Policies for Reinforcement Learningby Hector Kohler, Quentin Delfosse, Riad Akrour,…
SFDDM: Single-fold Distillation for Diffusion modelsby Chi Hong, Jiyue Huang, Robert Birke, Dick Epema, Stefanie…
Recurrent Early Exits for Federated Learning with Heterogeneous Clientsby Royson Lee, Javier Fernandez-Marques, Shell Xu…
GIFT: Unlocking Full Potential of Labels in Distilled Dataset at Near-zero Costby Xinyi Shang, Peng…