An Efficient Newton Method for Extreme Similarity Learning with Nonlinear Embeddings

10/26/2020
by   Bowen Yuan, et al.
26

We study the problem of learning similarity by using nonlinear embedding models (e.g., neural networks) from all possible pairs. This problem is well-known for its difficulty of training with the extreme number of pairs. Existing optimization methods extended from stochastic gradient methods suffer from slow convergence and high complexity per pass of all possible pairs. Inspired by some recent works reporting that Newton methods are competitive for training certain types of neural networks, in this work, we novelly apply the Newton method for this problem. A prohibitive cost depending on the extreme number of pairs occurs if the Newton method is directly applied. We propose an efficient algorithm which successfully eliminates the cost. Our proposed algorithm can take advantage of second-order information and lower time complexity per pass of all possible pairs. Experiments conducted on large-scale data sets demonstrate that the proposed algorithm is more efficient than existing algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/14/2018

Newton Methods for Convolutional Neural Networks

Deep learning involves a difficult non-convex optimization problem, whic...
research
04/06/2020

Deep Neural Network Learning with Second-Order Optimizers – a Practical Study with a Stochastic Quasi-Gauss-Newton Method

Training in supervised deep learning is computationally demanding, and t...
research
12/14/2021

SC-Reg: Training Overparameterized Neural Networks under Self-Concordant Regularization

In this paper we propose the SC-Reg (self-concordant regularization) fra...
research
07/18/2018

Efficient Training on Very Large Corpora via Gramian Estimation

We study the problem of learning similarity functions over very large co...
research
02/01/2018

Distributed Newton Methods for Deep Neural Networks

Deep learning involves a difficult non-convex optimization problem with ...
research
11/04/2015

adaQN: An Adaptive Quasi-Newton Algorithm for Training RNNs

Recurrent Neural Networks (RNNs) are powerful models that achieve except...
research
08/17/2023

Dual Gauss-Newton Directions for Deep Learning

Inspired by Gauss-Newton-like methods, we study the benefit of leveragin...

Please sign up or login with your details

Forgot password? Click here to reset