Extrapolation in NLP

05/17/2018
by   Jeff Mitchell, et al.
0

We argue that extrapolation to examples outside the training space will often be easier for models that capture global structures, rather than just maximise their local fit to the training data. We show that this is true for two popular models: the Decomposable Attention Model and word2vec.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset