A Discontinuous Neural Network for Non-Negative Sparse Approximation

03/21/2016
by   Martijn Arts, et al.
0

This paper investigates a discontinuous neural network which is used as a model of the mammalian olfactory system and can more generally be applied to solve non-negative sparse approximation problems. By inherently limiting the systems integrators to having non-negative outputs, the system function becomes discontinuous since the integrators switch between being inactive and being active. It is shown that the presented network converges to equilibrium points which are solutions to general non-negative least squares optimization problems. We specify a Caratheodory solution and prove that the network is stable, provided that the system matrix has full column-rank. Under a mild condition on the equilibrium point, we show that the network converges to its equilibrium within a finite number of switches. Two applications of the neural network are shown. Firstly, we apply the network as a model of the olfactory system and show that in principle it may be capable of performing complex sparse signal recovery tasks. Secondly, we generalize the application to include non-negative sparse approximation problems and compare the recovery performance to a classical non-negative basis pursuit denoising algorithm. We conclude that the recovery performance differs only marginally from the classical algorithm, while the neural network has the advantage that no performance critical regularization parameter has to be chosen prior to recovery.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/28/2020

DeepMP for Non-Negative Sparse Decomposition

Non-negative signals form an important class of sparse signals. Many alg...
research
01/17/2019

Sparse Non-Negative Recovery from Biased Subgaussian Measurements using NNLS

We investigate non-negative least squares (NNLS) for the recovery of spa...
research
03/12/2013

Recovering Non-negative and Combined Sparse Representations

The non-negative solution to an underdetermined linear system can be uni...
research
05/04/2012

Non-negative least squares for high-dimensional linear models: consistency and sparse recovery without regularization

Least squares fitting is in general not useful for high-dimensional line...
research
07/18/2022

Non-negative Least Squares via Overparametrization

In many applications, solutions of numerical problems are required to be...
research
07/20/2022

Fixed Points of Cone Mapping with the Application to Neural Networks

We derive conditions for the existence of fixed points of cone mappings ...
research
09/08/2017

Applications of an algorithm for solving Fredholm equations of the first kind

In this paper we use an iterative algorithm for solving Fredholm equatio...

Please sign up or login with your details

Forgot password? Click here to reset