Benefits and Pitfalls of Reinforcement Learning for Language Model Planning: A Theoretical Perspective

Agentic AI
Published: arXiv: 2509.22613v1
Authors

Siwei Wang Yifei Shen Haoran Sun Shi Feng Shang-Hua Teng Li Dong Yaru Hao Wei Chen

Abstract

Recent reinforcement learning (RL) methods have substantially enhanced the planning capabilities of Large Language Models (LLMs), yet the theoretical basis for their effectiveness remains elusive. In this work, we investigate RL's benefits and limitations through a tractable graph-based abstraction, focusing on policy gradient (PG) and Q-learning methods. Our theoretical analyses reveal that supervised fine-tuning (SFT) may introduce co-occurrence-based spurious solutions, whereas RL achieves correct planning primarily through exploration, underscoring exploration's role in enabling better generalization. However, we also show that PG suffers from diversity collapse, where output diversity decreases during training and persists even after perfect accuracy is attained. By contrast, Q-learning provides two key advantages: off-policy learning and diversity preservation at convergence. We further demonstrate that careful reward design is necessary to prevent reward hacking in Q-learning. Finally, applying our framework to the real-world planning benchmark Blocksworld, we confirm that these behaviors manifest in practice.

Paper Summary

Problem
Large Language Models (LLMs) have made significant progress in planning capabilities, but the theoretical basis for their effectiveness remains unclear. Researchers are trying to understand how reinforcement learning (RL) methods enhance planning in LLMs, and what are the benefits and limitations of using RL in this context.
Key Innovation
The paper presents a theoretical analysis of RL-based planning in LLMs, focusing on policy gradient and Q-learning methods. The authors develop a graph-based abstraction to investigate the learning dynamics of RL-based planning and compare it with supervised fine-tuning (SFT). The key innovation lies in the identification of exploration's role in achieving better generalization and the demonstration of Q-learning's advantages, including off-policy learning and diversity preservation.
Practical Impact
The research has practical implications for the development of more robust, scalable, and generalizable planning systems in LLMs. By understanding the benefits and limitations of RL-based planning, researchers can design more effective planning frameworks that leverage exploration to achieve better generalization. This can lead to improved performance in various applications, such as task decomposition, visual-language spatial navigation, and long-horizon robotics tasks.
Analogy / Intuitive Explanation
Imagine you're trying to solve a complex puzzle, and you have two different approaches: one that relies on memorization (SFT) and another that uses exploration and learning (RL). The SFT approach is like trying to find the solution by looking at a map of the puzzle, whereas the RL approach is like exploring the puzzle itself, learning from your mistakes, and adapting your strategy to find the solution. The paper shows that the RL approach is more effective in achieving better generalization, but it also has its own limitations, such as diversity collapse.
Paper Information
Categories:
cs.AI cs.LG stat.ML
Published Date:

arXiv ID:

2509.22613v1

Quick Actions