
-
Don't Fix What ain't Broke: Near-optimal Local Convergence of Alternating Gradient Descent-Ascent for Minimax Optimization
Minimax optimization has recently gained a lot of attention as adversari...
read it
-
Computational frameworks for homogenization and multiscale stability analyses of nonlinear periodic metamaterials
This paper presents a consistent computational framework for multiscale ...
read it
-
A Unified Analysis of First-Order Methods for Smooth Games via Integral Quadratic Constraints
The theory of integral quadratic constraints (IQCs) allows the certifica...
read it
-
On the Suboptimality of Negative Momentum for Minimax Optimization
Smooth game optimization has recently attracted great interest in machin...
read it
-
Picking Winning Tickets Before Training by Preserving Gradient Flow
Overparameterization has been shown to benefit both the optimization and...
read it
-
On Solving Minimax Optimization Locally: A Follow-the-Ridge Approach
Many tasks in modern machine learning can be formulated as finding equil...
read it
-
Which Algorithmic Choices Matter at Which Batch Sizes? Insights From a Noisy Quadratic Model
Increasing the batch size is a popular way to speed up neural network tr...
read it
-
Benchmarking Model-Based Reinforcement Learning
Model-based reinforcement learning (MBRL) is widely seen as having the p...
read it
-
Fast Convergence of Natural Gradient Descent for Overparameterized Neural Networks
Natural gradient descent has proven effective at mitigating the effects ...
read it
-
EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis
Reducing the test time resource requirements of a neural network while p...
read it
-
Computational Design of Finite Strain Auxetic Metamaterials via Topology Optimization and Nonlinear Homogenization
A novel computational framework for designing metamaterials with negativ...
read it
-
Functional Variational Bayesian Neural Networks
Variational Bayesian neural networks (BNNs) perform variational inferenc...
read it
-
Interplay Between Optimization and Generalization of Stochastic Gradient Descent with Covariance Noise
The choice of batch-size in a stochastic optimization algorithm plays a ...
read it
-
Eigenvalue Corrected Noisy Natural Gradient
Variational Bayesian neural networks combine the flexibility of deep lea...
read it
-
Three Mechanisms of Weight Decay Regularization
Weight decay is one of the standard tricks in the neural network toolbox...
read it
-
Differentiable Compositional Kernel Learning for Gaussian Processes
The generalization properties of Gaussian processes depend heavily on th...
read it
-
Noisy Natural Gradient as Variational Inference
Combining the flexibility of deep learning with Bayesian uncertainty est...
read it
-
Deformable Convolutional Networks
Convolutional neural networks (CNNs) are inherently limited to model geo...
read it