Ada-QPacknet – adaptive pruning with bit width reduction as an efficient continual learning method without forgetting

08/14/2023
by   Marcin Pietroń, et al.
0

Continual Learning (CL) is a process in which there is still huge gap between human and deep learning model efficiency. Recently, many CL algorithms were designed. Most of them have many problems with learning in dynamic and complex environments. In this work new architecture based approach Ada-QPacknet is described. It incorporates the pruning for extracting the sub-network for each task. The crucial aspect in architecture based CL methods is theirs capacity. In presented method the size of the model is reduced by efficient linear and nonlinear quantisation approach. The method reduces the bit-width of the weights format. The presented results shows that hybrid 8 and 4-bit quantisation achieves similar accuracy as floating-point sub-network on a well-know CL scenarios. To our knowledge it is the first CL strategy which incorporates both compression techniques pruning and quantisation for generating task sub-networks. The presented algorithm was tested on well-known episode combinations and compared with most popular algorithms. Results show that proposed approach outperforms most of the CL strategies in task and class incremental scenarios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/15/2019

Compacting, Picking and Growing for Unforgetting Continual Learning

Continual lifelong learning is essential to many applications. In this p...
research
10/28/2020

A Study on Efficiency in Continual Learning Inspired by Human Learning

Humans are efficient continual learning systems; we continually learn ne...
research
03/27/2023

Forget-free Continual Learning with Soft-Winning SubNetworks

Inspired by Regularized Lottery Ticket Hypothesis (RLTH), which states t...
research
03/17/2022

Continual Learning Based on OOD Detection and Task Masking

Existing continual learning techniques focus on either task incremental ...
research
04/08/2020

Continual Learning with Gated Incremental Memories for sequential data processing

The ability to learn in dynamic, nonstationary environments without forg...
research
07/23/2019

Adaptive Compression-based Lifelong Learning

The problem of a deep learning model losing performance on a previously ...
research
07/11/2023

Domain-Agnostic Neural Architecture for Class Incremental Continual Learning in Document Processing Platform

Production deployments in complex systems require ML architectures to be...

Please sign up or login with your details

Forgot password? Click here to reset