While large language models (LLMs) exhibit impressive language understan...
Despite the advancements of open-source large language models (LLMs) and...
As large language models (LLMs) generate texts with increasing fluency a...
Parameter-efficient tuning (PET) methods can effectively drive extremely...
This work examines the presence of modularity in pre-trained Transformer...
Injecting external knowledge can improve the performance of pre-trained
...
Parameter-efficient tuning methods (PETs) have achieved promising result...
Large-scale pre-trained models (PTMs) have been widely used in
document-...
Continual pre-training is the paradigm where pre-trained language models...
Long-form question answering (LFQA) aims at answering complex, open-ende...
Recent research demonstrates that external knowledge injection can advan...
Humans possess an extraordinary ability to create and utilize tools, all...
Federated Learning has become a widely-used framework which allows learn...
The diverse relationships among real-world events, including coreference...
Recent years have witnessed the prevalent application of pre-trained lan...
Delta tuning (DET, also known as parameter-efficient tuning) is deemed a...
Even though the large-scale language models have achieved excellent
perf...
Knowledge graphs, as the cornerstone of many AI applications, usually fa...
Investigating better ways to reuse the released pre-trained language mod...
Prompting, which casts downstream applications as language modeling task...
How to build and use dialogue data efficiently, and how to deploy models...
Adaptive learning aims to stimulate and meet the needs of individual
lea...
Existing reference-free metrics have obvious limitations for evaluating
...
Current pre-trained language models (PLM) are typically trained with sta...
Pre-trained language models (PLMs) cannot well recall rich factual knowl...
As many fine-tuned pre-trained language models (PLMs) with promising
per...
Prompt tuning (PT) is a promising parameter-efficient method to utilize
...
How can pre-trained language models (PLMs) learn universal representatio...
Backdoor attacks, which maliciously control a well-trained model's outpu...
The class imbalance problem, as an important issue in learning node
repr...
Transformer-based pre-trained language models can achieve superior
perfo...
Knowledge distillation (KD) has been proved effective for compressing
la...
Named Entity Recognition (NER) and Relation Extraction (RE) are the core...
Hyperbolic neural networks have shown great potential for modeling compl...
Event extraction (EE) has considerably benefited from pre-trained langua...
Recent explorations of large-scale pre-trained language models (PLMs) su...
Existing pre-trained language models (PLMs) are often computationally
ex...
Distantly supervised (DS) relation extraction (RE) has attracted much
at...
Fine-tuning pre-trained language models (PLMs) has demonstrated its
effe...
This book aims to review and present the recent advances of distributed
...
Pre-trained Language Models (PLMs) have shown strong performance in vari...
Dynamic early exiting aims to accelerate pre-trained language models' (P...
Knowledge graph embedding (KGE), aiming to embed entities and relations ...
Graph embedding (GE) methods embed nodes (and/or edges) in graph into a
...
Neural models have achieved remarkable success on relation extraction (R...
Several recent efforts have been devoted to enhancing pre-trained langua...
Non-autoregressive neural machine translation (NAT) predicts the entire
...
Event detection (ED), which identifies event trigger words and classifie...
Language representation models such as BERT could effectively capture
co...
Relational facts are an important component of human knowledge, which ar...