A Preliminary Evaluation of ChatGPT in Requirements Information Retrieval

04/25/2023
by   Jianzhang Zhang, et al.
0

Context: Recently, many illustrative examples have shown ChatGPT's impressive ability to perform programming tasks and answer general domain questions. Objective: We empirically evaluate how ChatGPT performs on requirements analysis tasks to derive insights into how generative large language model, represented by ChatGPT, influence the research and practice of natural language processing for requirements engineering. Method: We design an evaluation pipeline including two common requirements information retrieval tasks, four public datasets involving two typical requirements artifacts, querying ChatGPT with fixed task prompts, and quantitative and qualitative results analysis. Results: Quantitative results show that ChatGPT achieves comparable or better Fβ values in all datasets under a zero-shot setting. Qualitative analysis further illustrates ChatGPT's powerful natural language processing ability and limited requirements engineering domain knowledge. Conclusion: The evaluation results demonstrate ChatGPT' impressive ability to retrieve requirements information from different types artifacts involving multiple languages under a zero-shot setting. It is worthy for the research and industry communities to study generative large language model based requirements retrieval models and to develop corresponding tools.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/02/2022

Pop Quiz! Can a Large Language Model Help With Reverse Engineering?

Large language models (such as OpenAI's Codex) have demonstrated impress...
research
05/26/2023

Zero is Not Hero Yet: Benchmarking Zero-Shot Performance of LLMs for Financial Tasks

Recently large language models (LLMs) like ChatGPT have shown impressive...
research
08/04/2023

ChatGPT for GTFS: From Words to Information

The General Transit Feed Specification (GTFS) standard for publishing tr...
research
09/14/2023

An Empirical Evaluation of Prompting Strategies for Large Language Models in Zero-Shot Clinical Natural Language Processing

Large language models (LLMs) have shown remarkable capabilities in Natur...
research
11/16/2022

Technical Report on Neural Language Models and Few-Shot Learning for Systematic Requirements Processing in MDSE

Systems engineering, in particular in the automotive domain, needs to co...
research
05/04/2020

On Systematically Building a Controlled Natural Language for Functional Requirements

[Context] Natural language (NL) is pervasive in software requirements sp...
research
05/29/2022

Decoupling Knowledge from Memorization: Retrieval-augmented Prompt Learning

Prompt learning approaches have made waves in natural language processin...

Please sign up or login with your details

Forgot password? Click here to reset