Learning Admissible Heuristics for A*: Theory and Practice

Agentic AI
Published: arXiv: 2509.22626v1
Authors

Ehsan Futuhi Nathan R. Sturtevant

Abstract

Heuristic functions are central to the performance of search algorithms such as A-star, where admissibility - the property of never overestimating the true shortest-path cost - guarantees solution optimality. Recent deep learning approaches often disregard admissibility and provide limited guarantees on generalization beyond the training data. This paper addresses both of these limitations. First, we pose heuristic learning as a constrained optimization problem and introduce Cross-Entropy Admissibility (CEA), a loss function that enforces admissibility during training. On the Rubik's Cube domain, this method yields near-admissible heuristics with significantly stronger guidance than compressed pattern database (PDB) heuristics. Theoretically, we study the sample complexity of learning heuristics. By leveraging PDB abstractions and the structural properties of graphs such as the Rubik's Cube, we tighten the bound on the number of training samples needed for A-star to generalize. Replacing a general hypothesis class with a ReLU neural network gives bounds that depend primarily on the network's width and depth, rather than on graph size. Using the same network, we also provide the first generalization guarantees for goal-dependent heuristics.

Paper Summary

Problem
The main problem this research paper addresses is the challenge of learning admissible heuristics for A*, a widely used search algorithm. Admissible heuristics are crucial for A* to guarantee optimality, but traditional methods for computing them are often difficult to design and require significant domain knowledge. Recent deep learning approaches have shown promise but often disregard admissibility and provide limited guarantees on generalization.
Key Innovation
The key innovation of this paper is the introduction of Cross-Entropy Admissibility (CEA), a novel loss function that enforces admissibility during training. CEA is a constrained optimization problem that balances admissibility and heuristic strength, allowing the model to learn effective and admissible heuristics. The paper also provides theoretical guarantees on the sample complexity of learning heuristics, showing that the number of training samples needed for A* to generalize can be significantly reduced by leveraging PDB abstractions and graph structure.
Practical Impact
This research has significant practical impact on various applications that rely on A* and other heuristic search algorithms. By learning admissible heuristics, the paper demonstrates that it is possible to achieve near-optimal performance in practice while maintaining theoretical guarantees. The proposed framework can be applied to a wide range of domains, including robotics, logistics, and video games, where A* is commonly used. The paper also provides a foundation for future work on adapting both search and learning to work together while providing solution quality guarantees.
Analogy / Intuitive Explanation
Imagine you're trying to find the shortest path from your home to a new restaurant in an unfamiliar city. A* is like a map that helps you navigate through the city, but it needs a good estimate of the distance to the restaurant to guarantee you'll find the shortest path. Admissible heuristics are like a compass that always points in the right direction, never overestimating the distance. The paper proposes a new way to train this compass using deep learning, ensuring that it's always pointing in the right direction while also being effective in practice.
Paper Information
Categories:
cs.LG cs.AI
Published Date:

arXiv ID:

2509.22626v1

Quick Actions