Summary of Online Merging Optimizers For Boosting Rewards and Mitigating Tax in Alignment, by Keming Lu et al.
Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignmentby Keming Lu, Bowen Yu,…
Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignmentby Keming Lu, Bowen Yu,…
MMDisCo: Multi-Modal Discriminator-Guided Cooperative Diffusion for Joint Audio and Video Generationby Akio Hayakawa, Masato Ishii,…
AI Alignment with Changing and Influenceable Reward Functionsby Micah Carroll, Davis Foote, Anand Siththaranjan, Stuart…
Overcoming Negative Transfer by Online Selection: Distant Domain Adaptation for Fault Diagnosisby Ziyan Wang, Mohamed…
How Do the Architecture and Optimizer Affect Representation Learning? On the Training Dynamics of Representations…
ReMoDetect: Reward Models Recognize Aligned LLM’s Generationsby Hyunseok Lee, Jihoon Tack, Jinwoo ShinFirst submitted to…
Navigating the Safety Landscape: Measuring Risks in Finetuning Large Language Modelsby ShengYun Peng, Pin-Yu Chen,…
BeamVQ: Aligning Space-Time Forecasting Model via Self-training on Physics-aware Metricsby Hao Wu, Xingjian Shi, Ziyue…
Improving Data-aware and Parameter-aware Robustness for Continual Learningby Hanxi Xiao, Fan LyuFirst submitted to arxiv…
Effective Layer Pruning Through Similarity Metric Perspectiveby Ian Pons, Bruno Yamamoto, Anna H. Reali Costa,…