Summary of Uncovering Latent Human Wellbeing in Language Model Embeddings, by Pedro Freire et al.
Uncovering Latent Human Wellbeing in Language Model Embeddingsby Pedro Freire, ChengCheng Tan, Adam Gleave, Dan…
Uncovering Latent Human Wellbeing in Language Model Embeddingsby Pedro Freire, ChengCheng Tan, Adam Gleave, Dan…
ZeroG: Investigating Cross-dataset Zero-shot Transferability in Graphsby Yuhan Li, Peisong Wang, Zhixun Li, Jeffrey Xu…
CultureLLM: Incorporating Cultural Differences into Large Language Modelsby Cheng Li, Mengzhou Chen, Jindong Wang, Sunayana…
The Unreasonable Effectiveness of Eccentric Automatic Promptsby Rick Battle, Teja GollapudiFirst submitted to arxiv on:…
Active Preference Optimization for Sample Efficient RLHFby Nirjhar Das, Souradip Chakraborty, Aldo Pacchiano, Sayak Ray…
Subgraph-level Universal Prompt Tuningby Junhyun Lee, Wooseong Yang, Jaewoo KangFirst submitted to arxiv on: 16…
Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustmentby Rui Yang, Xiaoman Pan, Feng…
Crafting a Good Prompt or Providing Exemplary Dialogues? A Study of In-Context Learning for Persona-based…
RS-DPO: A Hybrid Rejection Sampling and Direct Preference Optimization Method for Alignment of Large Language…
Efficient Prompt Optimization Through the Lens of Best Arm Identificationby Chengshuai Shi, Kun Yang, Zihan…