Optimal Size-Performance Tradeoffs: Weighing PoS Tagger Models

04/16/2021
by   Magnus Jacobsen, et al.
0

Improvement in machine learning-based NLP performance are often presented with bigger models and more complex code. This presents a trade-off: better scores come at the cost of larger tools; bigger models tend to require more during training and inference time. We present multiple methods for measuring the size of a model, and for comparing this with the model's performance. In a case study over part-of-speech tagging, we then apply these techniques to taggers for eight languages and present a novel analysis identifying which taggers are size-performance optimal. Results indicate that some classical taggers place on the size-performance skyline across languages. Further, although the deep models have highest performance for multiple scores, it is often not the most complex of these that reach peak performance.

READ FULL TEXT
research
10/17/2021

Predicting the Performance of Multilingual NLP Models

Recent advancements in NLP have given us models like mBERT and XLMR that...
research
10/31/2016

Experiments with POS Tagging Code-mixed Indian Social Media Text

This paper presents Centre for Development of Advanced Computing Mumbai'...
research
03/22/2019

Data Augmentation via Dependency Tree Morphing for Low-Resource Languages

Neural NLP systems achieve high scores in the presence of sizable traini...
research
04/29/2020

A Cross-Genre Ensemble Approach to Robust Reddit Part of Speech Tagging

Part of speech tagging is a fundamental NLP task often regarded as solve...
research
06/03/2021

How to Adapt Your Pretrained Multilingual Model to 1600 Languages

Pretrained multilingual models (PMMs) enable zero-shot learning via cros...
research
07/11/2020

Deep or Simple Models for Semantic Tagging? It Depends on your Data [Experiments]

Semantic tagging, which has extensive applications in text mining, predi...

Please sign up or login with your details

Forgot password? Click here to reset