NukeLM: Pre-Trained and Fine-Tuned Language Models for the Nuclear and Energy Domains

05/25/2021
by   Lee Burke, et al.
0

Natural language processing (NLP) tasks (text classification, named entity recognition, etc.) have seen revolutionary improvements over the last few years. This is due to language models such as BERT that achieve deep knowledge transfer by using a large pre-trained model, then fine-tuning the model on specific tasks. The BERT architecture has shown even better performance on domain-specific tasks when the model is pre-trained using domain-relevant texts. Inspired by these recent advancements, we have developed NukeLM, a nuclear-domain language model pre-trained on 1.5 million abstracts from the U.S. Department of Energy Office of Scientific and Technical Information (OSTI) database. This NukeLM model is then fine-tuned for the classification of research articles into either binary classes (related to the nuclear fuel cycle [NFC] or not) or multiple categories related to the subject of the article. We show that continued pre-training of a BERT-style architecture prior to fine-tuning yields greater performance on both article classification tasks. This information is critical for properly triaging manuscripts, a necessary task for better understanding citation networks that publish in the nuclear space, and for uncovering new areas of research in the nuclear (or nuclear-relevant) domains.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/04/2022

A Comparison of SVM against Pre-trained Language Models (PLMs) for Text Classification Tasks

The emergence of pre-trained language models (PLMs) has shown great succ...
research
03/31/2022

ESGBERT: Language Model to Help with Classification Tasks Related to Companies Environmental, Social, and Governance Practices

Environmental, Social, and Governance (ESG) are non-financial factors th...
research
11/04/2021

CoreLM: Coreference-aware Language Model Fine-Tuning

Language Models are the underpin of all modern Natural Language Processi...
research
08/12/2022

Pre-training Tasks for User Intent Detection and Embedding Retrieval in E-commerce Search

BERT-style models pre-trained on the general corpus (e.g., Wikipedia) an...
research
07/15/2019

Myers-Briggs Personality Classification and Personality-Specific Language Generation Using Pre-trained Language Models

The Myers-Briggs Type Indicator (MBTI) is a popular personality metric t...
research
03/30/2020

NukeBERT: A Pre-trained language model for Low Resource Nuclear Domain

Significant advances have been made in recent years on Natural Language ...
research
02/25/2021

PharmKE: Knowledge Extraction Platform for Pharmaceutical Texts using Transfer Learning

The challenge of recognizing named entities in a given text has been a v...

Please sign up or login with your details

Forgot password? Click here to reset