Summary of Gdpo: Learning to Directly Align Language Models with Diversity Using Gflownets, by Oh Joon Kwon et al.
GDPO: Learning to Directly Align Language Models with Diversity Using GFlowNetsby Oh Joon Kwon, Daiki…
GDPO: Learning to Directly Align Language Models with Diversity Using GFlowNetsby Oh Joon Kwon, Daiki…
PopAlign: Diversifying Contrasting Patterns for a More Comprehensive Alignmentby Zekun Moore Wang, Shawn Wang, Kang…
Increasing the Difficulty of Automatically Generated Questions via Reinforcement Learning with Synthetic Preferenceby William Thorne,…
CodePMP: Scalable Preference Model Pretraining for Large Language Model Reasoningby Huimu Yu, Xing Wu, Weidong…
Seeing Eye to AI: Human Alignment via Gaze-Based Response Rewards for Large Language Modelsby Angela…
The Phenomenology of Machine: A Comprehensive Analysis of the Sentience of the OpenAI-o1 Model Integrating…
Elephant in the Room: Unveiling the Impact of Reward Model Quality in Alignmentby Yan Liu,…
Just Say What You Want: Only-prompting Self-rewarding Online Preference Optimizationby Ruijie Xu, Zhihan Liu, Yongfei…
Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedbackby Jiayi Zhou, Jiaming Ji, Juntao…
Minor SFT loss for LLM fine-tune to increase performance and reduce model deviationby Shiming Xie,…