Mapping Researcher Activity based on Publication Data by means of Transformers

06/15/2023
by   Zineddine Bettouche, et al.
0

Modern performance on several natural language processing (NLP) tasks has been enhanced thanks to the Transformer-based pre-trained language model BERT. We employ this concept to investigate a local publication database. Research papers are encoded and clustered to form a landscape view of the scientific topics, in which research is active. Authors working on similar topics can be identified by calculating the similarity between their papers. Based on this, we define a similarity metric between authors. Additionally we introduce the concept of self-similarity to indicate the topical variety of authors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2023

Large language models shape and are shaped by society: A survey of arXiv publication patterns

There has been a steep recent increase in the number of large language m...
research
04/14/2022

Brazilian Court Documents Clustered by Similarity Together Using Natural Language Processing Approaches with Transformers

Recent advances in Artificial intelligence (AI) have leveraged promising...
research
10/16/2019

NLPExplorer: Exploring the Universe of NLP Papers

Understanding the current research trends, problems, and their innovativ...
research
12/07/2021

A Scoping Review of Publicly Available Language Tasks in Clinical Natural Language Processing

Objective: to provide a scoping review of papers on clinical natural lan...
research
04/09/2021

Transformers: "The End of History" for NLP?

Recent advances in neural architectures, such as the Transformer, couple...
research
06/28/2018

Peerus Review: a tool for scientific experts finding

We propose a tool for experts finding applied to academic data generated...

Please sign up or login with your details

Forgot password? Click here to reset