Summary of Rose: a Reward-oriented Data Selection Framework For Llm Task-specific Instruction Tuning, by Yang Wu et al.
ROSE: A Reward-Oriented Data Selection Framework for LLM Task-Specific Instruction Tuningby Yang Wu, Huayi Zhang,…
ROSE: A Reward-Oriented Data Selection Framework for LLM Task-Specific Instruction Tuningby Yang Wu, Huayi Zhang,…
Unified Parameter-Efficient Unlearning for LLMsby Chenlu Ding, Jiancan Wu, Yancheng Yuan, Jinda Lu, Kai Zhang,…
On Foundation Models for Dynamical Systems from Purely Synthetic Databy Martin Ziegler, Andres Felipe Posada-Moreno,…
FonTS: Text Rendering with Typography and Style Controlsby Wenda Shi, Yiren Song, Dengming Zhang, Jiaming…
STEP: Enhancing Video-LLMs’ Compositional Reasoning by Spatio-Temporal Graph-guided Self-Trainingby Haiyi Qiu, Minghe Gao, Long Qian,…
Condense, Don’t Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruningby Mingyu Cao, Gen…
Energy-Efficient Split Learning for Fine-Tuning Large Language Models in Edge Networksby Zuguang Li, Shaohua Wu,…
Less is More: Efficient Model Merging with Binary Task Switchby Biqing Qi, Fangyuan Li, Zhen…
Reverse Thinking Makes LLMs Stronger Reasonersby Justin Chih-Yao Chen, Zifeng Wang, Hamid Palangi, Rujun Han,…
SURE-VQA: Systematic Understanding of Robustness Evaluation in Medical VQA Tasksby Kim-Celine Kahl, Selen Erkan, Jeremias…