NEZHA: Neural Contextualized Representation for Chinese Language Understanding

08/31/2019
by   Junqiu Wei, et al.
0

The pre-trained language models have achieved great successes in various natural language understanding (NLU) tasks due to its capacity to capture the deep contextualized information in text by pre-training on large-scale corpora. In this technical report, we present our practice of pre-training language models named NEZHA (NEural contextualiZed representation for CHinese lAnguage understanding) on Chinese corpora and finetuning for the Chinese NLU tasks. The current version of NEZHA is based on BERT with a collection of proven improvements, which include Functional Relative Positional Encoding as an effective positional encoding scheme, Whole Word Masking strategy, Mixed Precision Training and the LAMB Optimizer in training the models. The experimental results show that NEZHA achieves the state-of-the-art performances when finetuned on several representative Chinese tasks, including named entity recognition (People's Daily NER), sentence matching (LCQMC), Chinese sentiment classification (ChnSenti) and natural language inference (XNLI).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/19/2019

Pre-Training with Whole Word Masking for Chinese BERT

Bidirectional Encoder Representations from Transformers (BERT) has shown...
research
09/24/2020

AnchiBERT: A Pre-Trained Model for Ancient ChineseLanguage Understanding and Generation

Ancient Chinese is the essence of Chinese culture. There are several nat...
research
07/29/2019

ERNIE 2.0: A Continual Pre-training Framework for Language Understanding

Recently, pre-trained models have achieved state-of-the-art results in v...
research
06/15/2021

CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark

Artificial Intelligence (AI), along with the recent progress in biomedic...
research
10/11/2022

HUE: Pretrained Model and Dataset for Understanding Hanja Documents of Ancient Korea

Historical records in Korea before the 20th century were primarily writt...
research
04/19/2019

ERNIE: Enhanced Representation through Knowledge Integration

We present a novel language representation model enhanced by knowledge c...
research
07/05/2018

Chinese Lexical Analysis with Deep Bi-GRU-CRF Network

Lexical analysis is believed to be a crucial step towards natural langua...

Please sign up or login with your details

Forgot password? Click here to reset