Towards General-Purpose Model-Free Reinforcement Learning Paper • 2501.16142 • Published 18 days ago • 25
Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate Paper • 2501.17703 • Published 16 days ago • 53
Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach Paper • 2502.05171 • Published 7 days ago • 89
Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs Paper • 2501.18585 • Published 15 days ago • 52
GuardReasoner: Towards Reasoning-based LLM Safeguards Paper • 2501.18492 • Published 15 days ago • 81
The Differences Between Direct Alignment Algorithms are a Blur Paper • 2502.01237 • Published 11 days ago • 112
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training Paper • 2501.17161 • Published 17 days ago • 105
Reward-Guided Speculative Decoding for Efficient LLM Reasoning Paper • 2501.19324 • Published 14 days ago • 35
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models Paper • 2502.01061 • Published 11 days ago • 171
SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model Paper • 2502.02737 • Published 10 days ago • 161
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling Paper • 2502.06703 • Published 4 days ago • 116