
Learned Token Pruning for Transformers
A major challenge in deploying transformer models is their prohibitive i...
read it

QASR: Integeronly Zeroshot Quantization for Efficient Speech Recognition
Endtoend neural network models achieve improved performance on various...
read it

A Survey of Quantization Methods for Efficient Neural Network Inference
As soon as abstract mathematical computations were adapted to computatio...
read it

HessianAware Pruning and Optimal Neural Implant
Pruning is an effective method to reduce the memory footprint and FLOPs ...
read it

IBERT: Integeronly BERT Quantization
Transformer based models, like BERT and RoBERTa, have achieved stateof...
read it

HAWQV3: Dyadic Neural Network Quantization
Quantization is one of the key techniques used to make Neural Networks (...
read it

Boundary thickness and robustness in learning models
Robustness of machine learning models to various adversarial and nonadv...
read it

ADAHESSIAN: An Adaptive Second Order Optimizer for Machine Learning
We introduce AdaHessian, a second order stochastic optimization algorith...
read it

Rethinking Batch Normalization in Transformers
The standard normalization method for neural network (NN) models used in...
read it

ZeroQ: A Novel Zero Shot Quantization Framework
Quantization is a promising approach for reducing the inference time and...
read it

PyHessian: Neural Networks Through the Lens of the Hessian
We present PyHessian, a new scalable framework that enables fast computa...
read it

HAWQV2: Hessian Aware traceWeighted Quantization of Neural Networks
Quantization is an effective method for reducing memory footprint and in...
read it

Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization
Modern neural networks are increasingly bottlenecked by the limited capa...
read it

QBERT: Hessian Based Ultra Low Precision Quantization of BERT
Transformer based architectures have become defacto models used for a r...
read it

ANODEV2: A Coupled Neural ODE Evolution Framework
It has been observed that residual networks can be viewed as the explici...
read it

Inefficiency of KFAC for Large Batch Size Training
In stochastic optimization, large batch training can leverage parallel r...
read it

ANODE: Unconditionally Accurate MemoryEfficient Gradients for Neural ODEs
Residual neural networks can be viewed as the forward Euler discretizati...
read it

Trust Region Based Adversarial Attack on Neural Networks
Deep Neural Networks are quite vulnerable to adversarial perturbations. ...
read it

Parameter ReInitialization through Cyclical Batch Size Schedules
Optimal parameter initialization remains a crucial problem for neural ne...
read it

On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Increasing the minibatch size for stochastic gradient descent offers si...
read it

Identifying the Best Machine Learning Algorithms for Brain Tumor Segmentation, Progression Assessment, and Overall Survival Prediction in the BRATS Challenge
Gliomas are the most common primary brain malignancies, with different d...
read it

A Novel Domain Adaptation Framework for Medical Image Segmentation
We propose a segmentation framework that uses deep neural networks and i...
read it

Large batch size training of neural networks with adversarial training and secondorder information
Stochastic Gradient Descent (SGD) methods using randomly selected batche...
read it

CLAIRE: A distributedmemory solver for constrained large deformation diffeomorphic image registration
We introduce CLAIRE, a distributedmemory algorithm and software for sol...
read it

CoDesign of Deep Neural Nets and Neural Net Accelerators for Embedded Vision Applications
Deep Learning is arguably the most rapidly evolving research area in rec...
read it

SqueezeNext: HardwareAware Neural Network Design
One of the main barriers for deploying neural networks on embedded syste...
read it

PDEconstrained optimization in medical image analysis
PDEconstrained optimization problems find many applications in medical ...
read it

Hessianbased Analysis of Large Batch Training and Robustness to Adversaries
Large batch size training of Neural Networks has been shown to incur acc...
read it

Integrated Model, Batch and Domain Parallelism in Training Neural Networks
We propose a new integrated method of exploiting model, batch and domain...
read it

Integrated Model and Data Parallelism in Training Neural Networks
We propose a new integrated method of exploiting both model and data par...
read it

Distributedmemory large deformation diffeomorphic 3D image registration
We present a parallel distributedmemory algorithm for large deformation...
read it
Amir Gholami
is this you? claim profile