WPO: Enhancing RLHF with Weighted Preference Optimization Paper • 2406.11827 • Published Jun 17, 2024 • 17
view article Article Tokenization in Transformers v5: Simpler, Clearer, and More Modular +4 19 days ago • 96
Self-Play Preference Optimization for Language Model Alignment Paper • 2405.00675 • Published May 1, 2024 • 28
Noise Contrastive Alignment of Language Models with Explicit Rewards Paper • 2402.05369 • Published Feb 8, 2024 • 2
Towards Efficient and Exact Optimization of Language Model Alignment Paper • 2402.00856 • Published Feb 1, 2024 • 1
A General Theoretical Paradigm to Understand Learning from Human Preferences Paper • 2310.12036 • Published Oct 18, 2023 • 19
Provably Robust DPO: Aligning Language Models with Noisy Feedback Paper • 2403.00409 • Published Mar 1, 2024 • 2
Statistical Rejection Sampling Improves Preference Optimization Paper • 2309.06657 • Published Sep 13, 2023 • 14
Direct Preference Optimization: Your Language Model is Secretly a Reward Model Paper • 2305.18290 • Published May 29, 2023 • 64
DeepSeek-V3.2: Pushing the Frontier of Open Large Language Models Paper • 2512.02556 • Published Dec 2, 2025 • 244
Stabilizing Reinforcement Learning with LLMs: Formulation and Practices Paper • 2512.01374 • Published Dec 1, 2025 • 96
Go-Explore: a New Approach for Hard-Exploration Problems Paper • 1901.10995 • Published Jan 30, 2019 • 1
KTO: Model Alignment as Prospect Theoretic Optimization Paper • 2402.01306 • Published Feb 2, 2024 • 21