Support Vector Machines with the Hard-Margin Loss: Optimal Training via Combinatorial Benders' Cuts

07/15/2022
by   Ítalo Santana, et al.
0

The classical hinge-loss support vector machines (SVMs) model is sensitive to outlier observations due to the unboundedness of its loss function. To circumvent this issue, recent studies have focused on non-convex loss functions, such as the hard-margin loss, which associates a constant penalty to any misclassified or within-margin sample. Applying this loss function yields much-needed robustness for critical applications but it also leads to an NP-hard model that makes training difficult, since current exact optimization algorithms show limited scalability, whereas heuristics are not able to find high-quality solutions consistently. Against this background, we propose new integer programming strategies that significantly improve our ability to train the hard-margin SVM model to global optimality. We introduce an iterative sampling and decomposition approach, in which smaller subproblems are used to separate combinatorial Benders' cuts. Those cuts, used within a branch-and-cut algorithm, permit to converge much more quickly towards a global optimum. Through extensive numerical analyses on classical benchmark data sets, our solution algorithm solves, for the first time, 117 new data sets to optimality and achieves a reduction of 50 datasets of the benchmark.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2021

Learning with Smooth Hinge Losses

Due to the non-smoothness of the Hinge loss in SVM, it is difficult to o...
research
12/27/2018

Optimal Margin Distribution Network

Recent research about margin theory has proved that maximizing the minim...
research
03/16/2023

High-Dimensional Penalized Bernstein Support Vector Machines

The support vector machines (SVM) is a powerful classifier used for bina...
research
10/22/2019

Single Versus Union: Non-parallel Support Vector Machine Frameworks

Considering the classification problem, we summarize the nonparallel sup...
research
06/17/2020

Regularized ERM on random subspaces

We study a natural extension of classical empirical risk minimization, w...
research
05/16/2020

Classification vs regression in overparameterized regimes: Does the loss function matter?

We compare classification and regression tasks in the overparameterized ...
research
03/02/2020

Tropical Support Vector Machine and its Applications to Phylogenomics

Most data in genome-wide phylogenetic analysis (phylogenomics) is essent...

Please sign up or login with your details

Forgot password? Click here to reset