Improving Requirements Completeness: Automated Assistance through Large Language Models

08/03/2023
by   Dipeeka Luitel, et al.
0

Natural language (NL) is arguably the most prevalent medium for expressing systems and software requirements. Detecting incompleteness in NL requirements is a major challenge. One approach to identify incompleteness is to compare requirements with external sources. Given the rise of large language models (LLMs), an interesting question arises: Are LLMs useful external sources of knowledge for detecting potential incompleteness in NL requirements? This article explores this question by utilizing BERT. Specifically, we employ BERT's masked language model (MLM) to generate contextualized predictions for filling masked slots in requirements. To simulate incompleteness, we withhold content from the requirements and assess BERT's ability to predict terminology that is present in the withheld content but absent in the disclosed content. BERT can produce multiple predictions per mask. Our first contribution is determining the optimal number of predictions per mask, striking a balance between effectively identifying omissions in requirements and mitigating noise present in the predictions. Our second contribution involves designing a machine learning-based filter to post-process BERT's predictions and further reduce noise. We conduct an empirical evaluation using 40 requirements specifications from the PURE dataset. Our findings indicate that: (1) BERT's predictions effectively highlight terminology that is missing from requirements, (2) BERT outperforms simpler baselines in identifying relevant yet missing terminology, and (3) our filter significantly reduces noise in the predictions, enhancing BERT's effectiveness as a tool for completeness checking of requirements.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2023

Using Language Models for Enhancing the Completeness of Natural-language Requirements

[Context and motivation] Incompleteness in natural-language requirements...
research
05/13/2021

Are Larger Pretrained Language Models Uniformly Better? Comparing Performance at the Instance Level

Larger language models have higher accuracy on average, but are they bet...
research
02/09/2023

AI-based Question Answering Assistance for Analyzing Natural-language Requirements

By virtue of being prevalently written in natural language (NL), require...
research
01/20/2023

Phoneme-Level BERT for Enhanced Prosody of Text-to-Speech with Grapheme Predictions

Large-scale pre-trained language models have been shown to be helpful in...
research
08/14/2022

A Preliminary Study on the Potential Usefulness of Open Domain Model for Missing Software Requirements Recommendation

Completeness is one of the most important attributes of software require...
research
07/30/2023

User-Controlled Knowledge Fusion in Large Language Models: Balancing Creativity and Hallucination

In modern dialogue systems, the use of Large Language Models (LLMs) has ...

Please sign up or login with your details

Forgot password? Click here to reset