SALSA: Soup-based Alignment Learning for Stronger Adaptation in RLHF
Abstract
In Large Language Model (LLM) development, Reinforcement Learning from Human Feedback (RLHF) is crucial for aligning models with human values and preferences. RLHF traditionally relies on the Kullback-Leibler (KL) divergence between the current policy and a frozen initial policy as a reference, which is added as a penalty in policy optimization algorithms like Proximal Policy Optimization (PPO). While this constraint prevents models from deviating too far from the initial checkpoint, it limits exploration of the reward landscape, reducing the model's ability to discover higher-quality solutions. As a result, policy optimization is often trapped in a narrow region of the parameter space, leading to suboptimal alignment and performance. This paper presents SALSA (Soup-based Alignment Learning for Stronger Adaptation), a novel approach designed to overcome these limitations by creating a more flexible and better located reference model through weight-space averaging of two independent supervised fine-tuned (SFT) models. This model soup allows for larger deviation in KL divergence and exploring a promising region of the solution space without sacrificing stability. By leveraging this more robust reference model, SALSA fosters better exploration, achieving higher rewards and improving model robustness, out-of-distribution generalization, and performance. We validate the effectiveness of SALSA through extensive experiments on popular open models (Llama2-7B, Mistral-7B, and Gemma-2B) across various benchmarks (MT-Bench, Arena-Hard, UltraFeedback), where it consistently surpasses PPO by fostering deeper exploration and achieving superior alignment in LLMs.
Community
In this paper, we leverage model soup to enhance RLHF. Traditional PPO restricts optimization by keeping the model close to a single reference point, but SALSA introduces flexibility by exploring higher-reward regions in parameter space. Through our experiments, we showed practically that the regions around model soup are associated with higher rewards. This approach is both very easy to implement and highly effective, achieving superior performance over PPO across three benchmarks (Arena-Hard, MT-Bench, and UltraFeedback) and three models (Llama2-7B, Mistral-7B, and the smaller Gemma-2B).
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Towards Reliable Alignment: Uncertainty-aware RLHF (2024)
- $\alpha$-DPO: Adaptive Reward Margin is What Direct Preference Optimization Needs (2024)
- Policy Filtration in RLHF to Fine-Tune LLM for Code Generation (2024)
- Self-Evolved Reward Learning for LLMs (2024)
- Weak-to-Strong Preference Optimization: Stealing Reward from Weak Aligned Model (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper