Large Reasoning Models Learn Better Alignment from Flawed Thinking Paper • 2510.00938 • Published 28 days ago • 57
Controllable Safety Alignment: Inference-Time Adaptation to Diverse Safety Requirements Paper • 2410.08968 • Published Oct 11, 2024 • 14