Learning Two-Layer Residual Networks with Nonparametric Function Estimation by Convex Programming

08/17/2020
by   Zhunxuan Wang, et al.
0

We focus on learning a two-layer residual neural network with preactivation by ReLU (preReLU-TLRN): Suppose the input 𝐱 is from a distribution with support space ℝ^d and the ground-truth generative model is a preReLU-TLRN, given by 𝐲 = B^∗[(A^∗𝐱)^+ + 𝐱], where ground-truth network parameters A^∗∈ℝ^d× d is a nonnegative full-rank matrix and B^∗∈ℝ^m× d is full-rank with m ≥ d. We design layerwise objectives as functionals whose analytic minimizers sufficiently express the exact ground-truth network in terms of its parameters and nonlinearities. Following this objective landscape, learning a preReLU-TLRN from finite samples can be formulated as convex programming with nonparametric function estimation: For each layer, we first formulate the corresponding empirical risk minimization (ERM) as convex quadratic programming (QP), then we show the solution space of the QP can be equivalently determined by a set of linear inequalities, which can then be efficiently solved by linear programming (LP). Experiments show the robustness and sample efficiency of our methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/01/2017

Learning One-hidden-layer Neural Networks with Landscape Design

We consider the problem of learning a one-hidden-layer neural network: w...
research
10/16/2018

Learning Two-layer Neural Networks with Symmetric Inputs

We give a new algorithm for learning a two-layer neural network under a ...
research
07/15/2022

Blessing of Nonconvexity in Deep Linear Models: Depth Flattens the Optimization Landscape Around the True Solution

This work characterizes the effect of depth on the optimization landscap...
research
10/04/2022

Convex and Nonconvex Sublinear Regression with Application to Data-driven Learning of Reach Sets

We consider estimating a compact set from finite data by approximating t...
research
05/18/2021

Sharp Restricted Isometry Property Bounds for Low-rank Matrix Recovery Problems with Corrupted Measurements

In this paper, we study a general low-rank matrix recovery problem with ...
research
12/29/2020

Supermodularity and valid inequalities for quadratic optimization with indicators

We study the minimization of a rank-one quadratic with indicators and sh...
research
03/02/2020

Semiparametric Nonlinear Bipartite Graph Representation Learning with Provable Guarantees

Graph representation learning is a ubiquitous task in machine learning w...

Please sign up or login with your details

Forgot password? Click here to reset