Dynamic Self-Attention : Computing Attention over Words Dynamically for Sentence Embedding

08/22/2018
by   Deunsol Yoon, et al.
0

In this paper, we propose Dynamic Self-Attention (DSA), a new self-attention mechanism for sentence embedding. We design DSA by modifying dynamic routing in capsule network (Sabouretal.,2017) for natural language processing. DSA attends to informative words with a dynamic weight vector. We achieve new state-of-the-art results among sentence encoding methods in Stanford Natural Language Inference (SNLI) dataset with the least number of parameters, while showing comparative results in Stanford Sentiment Treebank (SST) dataset.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset