Maximizing Use-Case Specificity through Precision Model Tuning

12/29/2022
by   Pranjali Awasthi, et al.
0

Language models have become increasingly popular in recent years for tasks like information retrieval. As use-cases become oriented toward specific domains, fine-tuning becomes default for standard performance. To fine-tune these models for specific tasks and datasets, it is necessary to carefully tune the model's hyperparameters and training techniques. In this paper, we present an in-depth analysis of the performance of four transformer-based language models on the task of biomedical information retrieval. The models we consider are DeepMind's RETRO (7B parameters), GPT-J (6B parameters), GPT-3 (175B parameters), and BLOOM (176B parameters). We compare their performance on the basis of relevance, accuracy, and interpretability, using a large corpus of 480000 research papers on protein structure/function prediction as our dataset. Our findings suggest that smaller models, with <10B parameters and fine-tuned on domain-specific datasets, tend to outperform larger language models on highly specific questions in terms of accuracy, relevancy, and interpretability by a significant margin (+50 generally better results on broader prompts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2019

What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning

Pretrained transformer-based language models have achieved state of the ...
research
08/06/2023

Improving Domain-Specific Retrieval by NLI Fine-Tuning

The aim of this article is to investigate the fine-tuning potential of n...
research
01/01/2021

WARP: Word-level Adversarial ReProgramming

Transfer learning from pretrained language models recently became the do...
research
08/26/2021

Fine-tuning Pretrained Language Models with Label Attention for Explainable Biomedical Text Classification

The massive growth of digital biomedical data is making biomedical text ...
research
06/19/2023

Fine-tuning Large Enterprise Language Models via Ontological Reasoning

Large Language Models (LLMs) exploit fine-tuning as a technique to adapt...
research
05/26/2023

Improving accuracy of GPT-3/4 results on biomedical data using a retrieval-augmented language model

Large language models (LLMs) have made significant advancements in natur...
research
07/06/2019

Qwant Research @DEFT 2019: Document matching and information retrieval using clinical cases

This paper reports on Qwant Research contribution to tasks 2 and 3 of th...

Please sign up or login with your details

Forgot password? Click here to reset