Designing recurrent neural networks by unfolding an l1-l1 minimization algorithm

02/18/2019
by   Hung Duy Le, et al.
0

We propose a new deep recurrent neural network (RNN) architecture for sequential signal reconstruction. Our network is designed by unfolding the iterations of the proximal gradient method that solves the l1-l1 minimization problem. As such, our network leverages by design that signals have a sparse representation and that the difference between consecutive signal representations is also sparse. We evaluate the proposed model in the task of reconstructing video frames from compressive measurements and show that it outperforms several state-of-the-art RNN models.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset