In the past years, the application of neural networks as an alternative ...
Quantizing neural networks is one of the most effective methods for achi...
Neural network pruning and quantization techniques are almost as old as
...
The input tokens to Vision Transformers carry little semantic meaning as...
Transformer models have been widely adopted in various domains over the ...
Mixture of Experts (MoE) are rising in popularity as a means to train
ex...
Recently, the idea of using FP8 as a number format for neural network
tr...
Neural network quantization is frequently used to optimize model size,
l...
When quantizing neural networks for efficient inference, low-bit integer...
Though the state-of-the architectures for semantic segmentation, such as...
When training neural networks with simulated quantization, we observe th...
Current methods for pruning neural network weights iteratively apply
mag...
While neural networks have advanced the frontiers in many machine learni...
Transformer-based architectures have become the de-facto standard models...
While neural networks have advanced the frontiers in many applications, ...
This work presents DONNA (Distilling Optimal Neural Network Architecture...
We present a differentiable joint pruning and quantization (DJPQ) scheme...
We introduce Bayesian Bits, a practical method for joint mixed precision...
When quantizing neural networks, assigning each floating-point weight to...
Unlike ReLU, newer activation functions (like Swish, H-swish, Mish) that...
Convolutional Neural Networks experience catastrophic forgetting when
op...
This paper presents a novel differentiable method for unstructured weigh...
We analyze the effect of quantizing weights and activations of neural
ne...
The success of deep neural networks in many real-world applications is
l...
We present a method for gating deep-learning architectures on a fine-gra...
We introduce a data-free quantization method for deep neural networks th...
Neural network quantization has become an important research area due to...