A Structured Sparse Neural Network and Its Matrix Calculations Algorithm

Gradient descent optimizations and backpropagation are the most common methods for training neural networks, but they are computationally expensive for real time applications, need high memory resources, and are difficult to converge for many networks and large datasets. [Pseudo]inverse models for training neural network have emerged as powerful tools to overcome these issues. In order to effectively implement these methods, structured pruning maybe be applied to produce sparse neural networks. Although sparse neural networks are efficient in memory usage, most of their algorithms use the same fully loaded matrix calculation methods which are not efficient for sparse matrices. Tridiagonal matrices are one of the frequently used candidates for structuring neural networks, but they are not flexible enough to handle underfitting and overfitting problems as well as generalization properties. In this paper, we introduce a nonsymmetric, tridiagonal matrix with offdiagonal sparse entries and offset sub and super-diagonals as well algorithms for its [pseudo]inverse and determinant calculations. Traditional algorithms for matrix calculations, specifically inversion and determinant, of these forms are not efficient specially for large matrices, e.g. larger datasets or deeper networks. A decomposition for lower triangular matrices is developed and the original matrix is factorized into a set of matrices where their inverse matrices are calculated. For the cases where the matrix inverse does not exist, a least square type pseudoinverse is provided. The present method is a direct routine, i.e., executes in a predictable number of operations which is tested for randomly generated matrices with varying size. The results show significant improvement in computational costs specially when the size of matrix increases.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/28/2023

A Generalization of Continuous Relaxation in Structured Pruning

Deep learning harnesses massive parallel floating-point processing to tr...
research
09/16/2021

Sparse Factorization of Large Square Matrices

Square matrices appear in many machine learning problems and models. Opt...
research
10/30/2017

A Massively Parallel Algorithm for the Approximate Calculation of Inverse p-th Roots of Large Sparse Matrices

We present the submatrix method, a highly parallelizable method for the ...
research
09/23/2020

Analytical Inverse For The Symmetric Circulant Tridiagonal Matrix

Finding the inverse of a matrix is an open problem especially when it co...
research
12/07/2015

Gauss quadrature for matrix inverse forms with applications

We present a framework for accelerating a spectrum of machine learning a...
research
05/19/2021

Pseudo-Hadamard matrices of the first generation and an algorithm for producing them

Hadamard matrices in {0,1} presentation are square m× m matrices whose e...

Please sign up or login with your details

Forgot password? Click here to reset