Molding CNNs for text: non-linear, non-consecutive convolutions

08/17/2015
by   Tao Lei, et al.
0

The success of deep learning often derives from well-chosen operational building blocks. In this work, we revise the temporal convolution operation in CNNs to better adapt it to text processing. Instead of concatenating word representations, we appeal to tensor algebra and use low-rank n-gram tensors to directly exploit interactions between words already at the convolution stage. Moreover, we extend the n-gram convolution to non-consecutive words to recognize patterns with intervening words. Through a combination of low-rank tensors, and pattern weighting, we can efficiently evaluate the resulting convolution operation via dynamic programming. We test the resulting architecture on standard sentiment classification and news categorization tasks. Our model achieves state-of-the-art performance both in terms of accuracy and training speed. For instance, we obtain 51.2 fine-grained sentiment classification task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/12/2020

Seq2Tens: An Efficient Representation of Sequences by Low-Rank Tensor Projections

Sequential data such as time series, video, or text can be challenging t...
research
12/26/2013

Language Modeling with Power Low Rank Ensembles

We present power low rank ensembles (PLRE), a flexible framework for n-g...
research
04/02/2016

Embedding Lexical Features via Low-Rank Tensors

Modern NLP models rely heavily on engineered features, which often combi...
research
05/03/2019

Convolution is outer product

The inner product operation between tensors is the corner stone of deep ...
research
05/09/2019

Seesaw-Net: Convolution Neural Network With Uneven Group Convolution

In this paper, we are interested in boosting the representation capabili...
research
09/09/2023

Tensors Ranks and the Fine-Grained Complexity of Dynamic Programming

Generalizing work of Künnemann, Paturi, and Schneider [ICALP 2017], we s...
research
02/25/2021

Named Tensor Notation

We propose a notation for tensors with named axes, which relieves the au...

Please sign up or login with your details

Forgot password? Click here to reset