Cheap Orthogonal Constraints in Neural Networks: A Simple Parametrization of the Orthogonal and Unitary Group

01/24/2019
by   Mario Lezcano Casado, et al.
0

We introduce a novel approach to perform first-order optimization with orthogonal and unitary constraints. This approach is based on a parametrization stemming from Lie group theory through the exponential map. The parametrization transforms the constrained optimization problem into an unconstrained one over a Euclidean space, for which common first-order optimization methods can be used. The theoretical results presented are general enough to cover the special orthogonal group, the unitary group and, in general, any connected compact Lie group. We discuss how this and other parametrizations can be computed efficiently through an implementation trick, making numerically complex parametrizations usable at a negligible runtime cost in neural networks. In particular, we apply our results to RNNs with orthogonal recurrent weights, yielding a new architecture called expRNN. We demonstrate how our method constitutes a more robust approach to optimization with orthogonal constraints, showing faster, accurate, and more stable convergence in several tasks designed to test RNNs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2020

Stochastic Flows and Geometric Optimization on the Orthogonal Group

We present a new class of stochastic, geometrically-driven optimization ...
research
05/31/2023

A General Framework for Equivariant Neural Networks on Reductive Lie Groups

Reductive Lie Groups, such as the orthogonal groups, the Lorentz group, ...
research
04/18/2020

CWY Parametrization for Scalable Learning of Orthogonal and Stiefel Matrices

In this paper we propose a new approach for optimization over orthogonal...
research
05/26/2023

Bilipschitz group invariants

Consider the quotient of a real Hilbert space by a subgroup of its ortho...
research
04/27/2023

An Algorithm for Computing with Brauer's Group Equivariant Neural Network Layers

The learnable, linear neural network layers between tensor power spaces ...
research
01/26/2022

High-order symplectic Lie group methods on SO(n) using the polar decomposition

A variational integrator of arbitrarily high-order on the special orthog...
research
05/28/2019

Non-normal Recurrent Neural Network (nnRNN): learning long time dependencies while improving expressivity with transient dynamics

A recent strategy to circumvent the exploding and vanishing gradient pro...

Please sign up or login with your details

Forgot password? Click here to reset