Oracle Linguistic Graphs Complement a Pretrained Transformer Language Model: A Cross-formalism Comparison

12/15/2021
by   Jakob Prange, et al.
0

We examine the extent to which, in principle, linguistic graph representations can complement and improve neural language modeling. With an ensemble setup consisting of a pretrained Transformer and ground-truth graphs from one of 7 different formalisms, we find that, overall, semantic constituency structures are most useful to language modeling performance – outpacing syntactic constituency structures as well as syntactic and semantic dependency structures. Further, effects vary greatly depending on part-of-speech class. In sum, our findings point to promising tendencies in neuro-symbolic language modeling and invite future research quantifying the design choices made by different formalisms.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset