This paper focuses on addressing the practical yet challenging problem o...
Augmented Language Models (ALMs) empower large language models with the
...
Neural networks trained on distilled data often produce over-confident o...
Rare categories abound in a number of real-world networks and play a piv...
Spiking Transformers have gained considerable attention because they ach...
Biologically inspired Spiking Neural Networks (SNNs) have attracted
sign...
Various contrastive learning approaches have been proposed in recent yea...
Gradients have been exploited in proposal distributions to accelerate th...
Although sparse training has been successfully used in various
resource-...
Despite impressive performance on a wide variety of tasks, deep neural
n...
Over-parameterization of deep neural networks (DNNs) has shown high
pred...
Large-scale Transformer models bring significant improvements for variou...
Open domain question answering (ODQA) is a longstanding task aimed at
an...
Exploiting sparsity underlying neural networks has become one of the mos...
With the yearning for deep learning democratization, there are increasin...
Knowledge distillation (KD) methods compress large models into smaller
s...
Various graph contrastive learning models have been proposed to improve ...
Various pruning approaches have been proposed to reduce the footprint
re...
Transformer-based pre-trained language models have significantly improve...
Despite recent progress in Graph Neural Networks (GNNs), explaining
pred...
We consider the problem of learning predictive models from longitudinal ...
Selecting important variables and learning predictive models from
high-d...
We consider the problem of learning predictive models from longitudinal ...
Positive instance detection, especially for these in positive bags (true...