Semi-supervised knowledge distillation is a powerful training paradigm f...
One way of introducing sparsity into deep networks is by attaching an
ex...
It is well established that increasing scale in deep transformer network...
Distillation with unlabeled examples is a popular and powerful method fo...
Distilling knowledge from a large teacher model to a lightweight one is ...
Deep and wide neural networks successfully fit very complex functions to...
Graph neural networks have gained prominence due to their excellent
perf...
With the wide-spread availability of complex relational data, semi-super...
We develop an online learning algorithm for identifying unlabeled data p...
Neural network pruning is a popular technique used to reduce the inferen...
We present an efficient coreset construction algorithm for large-scale
S...
We present a provable, sampling-based approach for generating compact
Co...
We introduce a pruning algorithm that provably sparsifies the parameters...
The deployment of state-of-the-art neural networks containing millions o...