Summary of Combining Domain and Alignment Vectors to Achieve Better Knowledge-safety Trade-offs in Llms, by Megh Thakkar et al.
Combining Domain and Alignment Vectors to Achieve Better Knowledge-Safety Trade-offs in LLMsby Megh Thakkar, Yash…
Combining Domain and Alignment Vectors to Achieve Better Knowledge-Safety Trade-offs in LLMsby Megh Thakkar, Yash…
Aquila-plus: Prompt-Driven Visual-Language Models for Pixel-Level Remote Sensing Image Understandingby Kaixuan LuFirst submitted to arxiv…
IOPO: Empowering LLMs with Complex Instruction Following via Input-Output Preference Optimizationby Xinghua Zhang, Haiyang Yu,…
Aquila: A Hierarchically Aligned Visual-Language Model for Enhanced Remote Sensing Image Comprehensionby Kaixuan Lu, Ruiqian…
Benchmarking Distributional Alignment of Large Language Modelsby Nicole Meister, Carlos Guestrin, Tatsunori HashimotoFirst submitted to…
Rethinking Bradley-Terry Models in Preference-Based Reward Modeling: Foundations, Theory, and Alternativesby Hao Sun, Yunyi Shen,…
Bottom-Up and Top-Down Analysis of Values, Agendas, and Observations in Corpora and LLMsby Scott E.…
Combining Theory of Mind and Kindness for Self-Supervised Human-AI Alignmentby Joshua T. S. HewsonFirst submitted…
SMoA: Improving Multi-agent Large Language Models with Sparse Mixture-of-Agentsby Dawei Li, Zhen Tan, Peijia Qian,…
TODO: Enhancing LLM Alignment with Ternary Preferencesby Yuxiang Guo, Lu Yin, Bo Jiang, Jiaqi ZhangFirst…