Summary of Hard Prompts Made Interpretable: Sparse Entropy Regularization For Prompt Tuning with Rl, by Yunseon Choi et al.
Hard Prompts Made Interpretable: Sparse Entropy Regularization for Prompt Tuning with RLby Yunseon Choi, Sangmin…
Hard Prompts Made Interpretable: Sparse Entropy Regularization for Prompt Tuning with RLby Yunseon Choi, Sangmin…
Catastrophic Goodhart: regularizing RLHF with KL divergence does not mitigate heavy-tailed reward misspecificationby Thomas Kwa,…
OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learningby Yihang Yao, Zhepeng Cen, Wenhao Ding,…
On the Causal Sufficiency and Necessity of Multi-Modal Representation Learningby Jingyao Wang, Siyu Zhao, Wenwen…
DisenSemi: Semi-supervised Graph Classification via Disentangled Representation Learningby Yifan Wang, Xiao Luo, Chong Chen, Xian-Sheng…
BERTer: The Efficient Oneby Pradyumna Saligram, Andrew LanpouthakounFirst submitted to arxiv on: 19 Jul 2024CategoriesMain:…
Correcting the Mythos of KL-Regularization: Direct Alignment without Overoptimization via Chi-Squared Preference Optimizationby Audrey Huang,…
DropKAN: Regularizing KANs by masking post-activationsby Mohammed Ghaith AltarabichiFirst submitted to arxiv on: 17 Jul…
On the Calibration of Epistemic Uncertainty: Principles, Paradoxes and Conflictual Lossby Mohammed Fellaji, Frédéric Pennerath,…
Subject-driven Text-to-Image Generation via Preference-based Reinforcement Learningby Yanting Miao, William Loh, Suraj Kothawade, Pascal Poupart,…