Sample Complexity of Learning Heuristic Functions for Greedy-Best-First and A* Search

05/20/2022
by   Shinsaku Sakaue, et al.
0

Greedy best-first search (GBFS) and A* search (A*) are popular algorithms for path-finding on large graphs. Both use so-called heuristic functions, which estimate how close a vertex is to the goal. While heuristic functions have been handcrafted using domain knowledge, recent studies demonstrate that learning heuristic functions from data is effective in many applications. Motivated by this emerging approach, we study the sample complexity of learning heuristic functions for GBFS and A*. We build on a recent framework called data-driven algorithm design and evaluate the pseudo-dimension of a class of utility functions that measure the performance of parameterized algorithms. Assuming that a vertex set of size n is fixed, we present O(n n) and O(n^2 n) upper bounds on the pseudo-dimensions for GBFS and A*, respectively, parameterized by heuristic function values. The upper bound for A* can be improved to O(n^2 d) if every vertex has a degree of at most d and to O(n n) if edge weights are integers bounded by poly(n). We also give Ω(n) lower bounds for GBFS and A*, which imply that our bounds for GBFS and A* under the integer-weight condition are tight up to a n factor. Finally, we discuss a case where the performance of A* is measured by the suboptimality and show that we can sometimes obtain a better guarantee by combining a parameter-dependent worst-case bound with a sample complexity bound.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/11/2018

The Coin Problem in Constant Depth: Sample Complexity and Parity Gates

The δ-Coin Problem is the computational problem of distinguishing betwee...
research
09/11/2018

A Fixed-Depth Size-Hierarchy Theorem for AC^0[⊕] via the Coin Problem

We prove the first Fixed-depth Size-hierarchy Theorem for uniform AC^0[⊕...
research
01/03/2019

A modified greedy algorithm to improve bounds for the vertex cover number

In any attempt at designing an efficient algorithm for the minimum verte...
research
02/17/2020

Agnostic Q-learning with Function Approximation in Deterministic Systems: Tight Bounds on Approximation Error and Sample Complexity

The current paper studies the problem of agnostic Q-learning with functi...
research
04/21/2023

Prediction, Learning, Uniform Convergence, and Scale-sensitive Dimensions

We present a new general-purpose algorithm for learning classes of [0,1]...
research
05/28/2022

Provably Auditing Ordinary Least Squares in Low Dimensions

Measuring the stability of conclusions derived from Ordinary Least Squar...
research
06/21/2020

Refined bounds for algorithm configuration: The knife-edge of dual class approximability

Automating algorithm configuration is growing increasingly necessary as ...

Please sign up or login with your details

Forgot password? Click here to reset