Deep Language Networks: Joint Prompt Training of Stacked LLMs using Variational Inference Paper • 2306.12509 • Published Jun 21, 2023 • 14
The N+ Implementation Details of RLHF with PPO: A Case Study on TL;DR Summarization Paper • 2403.17031 • Published Mar 24, 2024 • 6
Generative Verifiers: Reward Modeling as Next-Token Prediction Paper • 2408.15240 • Published Aug 27, 2024 • 13
Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling Paper • 2408.16737 • Published Aug 29, 2024 • 1
Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models Paper • 2410.18252 • Published Oct 23, 2024 • 7
Putting the Value Back in RL: Better Test-Time Scaling by Unifying LLM Reasoners With Verifiers Paper • 2505.04842 • Published May 7 • 12
Multi-Turn Puzzles: Evaluating Interactive Reasoning and Strategic Dialogue in LLMs Paper • 2508.10142 • Published Aug 13 • 3
ReasoningMila/ServiceNowAI_R1_Distill_SFT_with_problems_and_responses Viewer • Updated May 22 • 1.68M • 107
ReasoningMila/ServiceNowAI_R1_Distill_SFT_with_problems_and_responses Viewer • Updated May 22 • 1.68M • 107
Putting the Value Back in RL: Better Test-Time Scaling by Unifying LLM Reasoners With Verifiers Paper • 2505.04842 • Published May 7 • 12
Leveraging recent advances in Pre-Trained Language Models forEye-Tracking Prediction Paper • 2110.04475 • Published Oct 9, 2021
When To Solve, When To Verify: Compute-Optimal Problem Solving and Generative Verification for LLM Reasoning Paper • 2504.01005 • Published Apr 1 • 15
ReasoningMila/syn_qs_and_soln_cleaned_0_and_less20_multiple_soln_per_qs_1937545 Viewer • Updated Mar 23 • 1.94M • 15