Refined bounds for algorithm configuration: The knife-edge of dual class approximability

06/21/2020
by   Maria-Florina Balcan, et al.
0

Automating algorithm configuration is growing increasingly necessary as algorithms come with more and more tunable parameters. It is common to tune parameters using machine learning, optimizing performance metrics such as runtime and solution quality. The training set consists of problem instances from the specific domain at hand. We investigate a fundamental question about these techniques: how large should the training set be to ensure that a parameter's average empirical performance over the training set is close to its expected, future performance? We answer this question for algorithm configuration problems that exhibit a widely-applicable structure: the algorithm's performance as a function of its parameters can be approximated by a "simple" function. We show that if this approximation holds under the L-infinity norm, we can provide strong sample complexity bounds. On the flip side, if the approximation holds only under the L-p norm for p smaller than infinity, it is not possible to provide meaningful sample complexity bounds in the worst case. We empirically evaluate our bounds in the context of integer programming, one of the most powerful tools in computer science. Via experiments, we obtain sample complexity bounds that are up to 700 times smaller than the previously best-known bounds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/18/2021

Improved Learning Bounds for Branch-and-Cut

Branch-and-cut is the most widely used algorithm for solving integer pro...
research
08/08/2019

How much data is sufficient to learn high-performing algorithms?

Algorithms for scientific analysis typically have tunable parameters tha...
research
12/24/2020

Generalization in portfolio-based algorithm selection

Portfolio-based algorithm selection has seen tremendous practical succes...
research
08/23/2022

Convergence bounds for nonlinear least squares for tensor recovery

We consider the problem of approximating a function in general nonlinear...
research
11/14/2016

Learning-Theoretic Foundations of Algorithm Configuration for Combinatorial Partitioning Problems

Max-cut, clustering, and many other partitioning problems that are of si...
research
05/20/2022

Sample Complexity of Learning Heuristic Functions for Greedy-Best-First and A* Search

Greedy best-first search (GBFS) and A* search (A*) are popular algorithm...
research
11/14/2020

Data-driven Algorithm Design

Data driven algorithm design is an important aspect of modern data scien...

Please sign up or login with your details

Forgot password? Click here to reset