As a popular channel pruning method for convolutional neural networks (C...
In this paper, we propose a feature affinity (FA) assisted knowledge
dis...
It has been shown by many researchers that transformers perform as well ...
In the last decade, convolutional neural networks (CNNs) have evolved to...
Differentiable architecture search (DARTS) is an effective method for
da...
To enable DNNs on edge devices like mobile phones, low-rank approximatio...
Deepening and widening convolutional neural networks (CNNs) significantl...
Following recent technological advances there is a growing interest in
b...
To accelerate DNNs inference, low-rank approximation has been widely ado...
Training activation quantized neural networks involves minimizing a piec...
ShuffleNet is a state-of-the-art light weight convolutional neural netwo...
Deploying a deep learning model on mobile/IoT devices is a challenging t...
The performance of Deep Neural Networks (DNNs) keeps elevating in recent...
Network quantization is an effective method for the deployment of neural...
Quantized deep neural networks (QDNNs) are attractive due to their much ...
We propose BinaryRelax, a simple two-phase algorithm, for training deep
...