Can LLMs Correct Themselves? A Benchmark of Self-Correction in LLMs Paper • 2510.16062 • Published 16 days ago • 1
BadVLA: Towards Backdoor Attacks on Vision-Language-Action Models via Objective-Decoupled Optimization Paper • 2505.16640 • Published May 22 • 3
Large Reasoning Models in Agent Scenarios: Exploring the Necessity of Reasoning Capabilities Paper • 2503.11074 • Published Mar 14 • 2
Vision Matters: Simple Visual Perturbations Can Boost Multimodal Math Reasoning Paper • 2506.09736 • Published Jun 11 • 9
Unsupervised Post-Training for Multi-Modal LLM Reasoning via GRPO Paper • 2505.22453 • Published May 28 • 46
Advancing Multimodal Reasoning via Reinforcement Learning with Cold Start Paper • 2505.22334 • Published May 28 • 36
Both Text and Images Leaked! A Systematic Analysis of Multimodal LLM Data Contamination Paper • 2411.03823 • Published Nov 6, 2024 • 49
Automating Safety Enhancement for LLM-based Agents with Synthetic Risk Scenarios Paper • 2505.17735 • Published May 23 • 3
NodeRAG: Structuring Graph-based RAG with Heterogeneous Nodes Paper • 2504.11544 • Published Apr 15 • 43
MLP-KAN: Unifying Deep Representation and Function Learning Paper • 2410.03027 • Published Oct 3, 2024 • 31