Enhancing Pre-trained Chinese Character Representation with Word-aligned Attention

11/07/2019
by   Yanzeng Li, et al.
0

Most Chinese pre-trained encoders take a character as a basic unit and learn representations according to character's external contexts, ignoring the semantics expressed in the word, which is the smallest meaningful unit in Chinese. Hence, we propose a novel word aligned attention to incorporate word segmentation information, which is complementary to various Chinese pre-trained language models. Specifically, we devise a mixed-pooling strategy to align the character level attention to the word level, and propose an effective fusion method to solve the potential issue of segmentation error propagation. As a result, word and character information are explicitly integrated at the fine-tuning procedure. Experimental results on various Chinese NLP benchmarks demonstrate that our model could bring another significant gain over several pre-trained models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/13/2022

Exploiting Word Semantics to Enrich Character Representations of Chinese Pre-trained Models

Most of the Chinese pre-trained models adopt characters as basic units f...
research
08/14/2023

A Novel Ehanced Move Recognition Algorithm Based on Pre-trained Models with Positional Embeddings

The recognition of abstracts is crucial for effectively locating the con...
research
05/23/2018

Enhancing Chinese Intent Classification by Dynamically Integrating Character Features into Word Embeddings with Ensemble Techniques

Intent classification has been widely researched on English data with de...
research
02/27/2018

A Hybrid Word-Character Model for Abstractive Summarization

Abstractive summarization is the popular research topic nowadays. Due to...
research
11/25/2019

Chinese Spelling Error Detection Using a Fusion Lattice LSTM

Spelling error detection serves as a crucial preprocessing in many natur...
research
09/08/2023

GLS-CSC: A Simple but Effective Strategy to Mitigate Chinese STM Models' Over-Reliance on Superficial Clue

Pre-trained models have achieved success in Chinese Short Text Matching ...
research
01/30/2023

PaCaNet: A Study on CycleGAN with Transfer Learning for Diversifying Fused Chinese Painting and Calligraphy

AI-Generated Content (AIGC) has recently gained a surge in popularity, p...

Please sign up or login with your details

Forgot password? Click here to reset