Efficient Ridge Solution for the Incremental Broad Learning System on Added Nodes by Inverse Cholesky Factorization of a Partitioned Matrix

11/12/2019
by   Hufei Zhu, et al.
0

To accelerate the existing Broad Learning System (BLS) for new added nodes in [7], we extend the inverse Cholesky factorization in [10] to deduce an efficient inverse Cholesky factorization for a Hermitian matrix partitioned into 2 * 2 blocks, which is utilized to develop the proposed BLS algorithm 1. The proposed BLS algorithm 1 compute the ridge solution (i.e, the output weights) from the inverse Cholesky factor of the Hermitian matrix in the ridge inverse, and update the inverse Cholesky factor efficiently. From the proposed BLS algorithm 1, we deduce the proposed ridge inverse, which can be obtained from the generalized inverse in [7] by just change one matrix in the equation to compute the newly added sub-matrix. We also modify the proposed algorithm 1 into the proposed algorithm 2, which is equivalent to the existing BLS algorithm [7] in terms of numerical computations. The proposed algorithms 1 and 2 can reduce the computational complexity, since usually the Hermitian matrix in the ridge inverse is smaller than the ridge inverse. With respect to the existing BLS algorithm, the proposed algorithms 1 and 2 usually require about 13 and 2 3 of complexities, respectively, while in numerical experiments they achieve the speedups (in each additional training time) of 2.40 - 2.91 and 1.36 - 1.60, respectively. Numerical experiments also show that the proposed algorithm 1 and the standard ridge solution always bear the same testing accuracy, and usually so do the proposed algorithm 2 and the existing BLS algorithm. The existing BLS assumes the ridge parameter lamda->0, since it is based on the generalized inverse with the ridge regression approximation. When the assumption of lamda-> 0 is not satisfied, the standard ridge solution obviously achieves a better testing accuracy than the existing BLS algorithm in numerical experiments.

READ FULL TEXT
research
05/14/2020

Efficient and Stable Algorithms to Extend Greville's Method to Partitioned Matrices Based on Inverse Cholesky Factorization

Greville's method has been utilized in (Broad Learn-ing System) BLS to p...
research
11/12/2019

Efficient Inverse-Free Algorithms for Extreme Learning Machine Based on the Recursive Matrix Inverse and the Inverse LDL' Factorization

The inverse-free extreme learning machine (ELM) algorithm proposed in [4...
research
05/21/2021

Low-Memory Implementations of Ridge Solutions for Broad Learning System with Incremental Learning

The existing low-memory BLS implementation proposed recently avoids the ...
research
10/17/2019

Reducing the Computational Complexity of Pseudoinverse for the Incremental Broad Learning System on Added Inputs

In this brief, we improve the Broad Learning System (BLS) [7] by reducin...
research
12/17/2013

The Matrix Ridge Approximation: Algorithms and Applications

We are concerned with an approximation problem for a symmetric positive ...
research
12/31/2019

Efficient Decremental Learning Algorithms for Broad Learning System

The decremented learning algorithms are required in machine learning, to...

Please sign up or login with your details

Forgot password? Click here to reset