Using Language Models for Enhancing the Completeness of Natural-language Requirements

02/09/2023
by   Dipeeka Luitel, et al.
0

[Context and motivation] Incompleteness in natural-language requirements is a challenging problem. [Question/problem] A common technique for detecting incompleteness in requirements is checking the requirements against external sources. With the emergence of language models such as BERT, an interesting question is whether language models are useful external sources for finding potential incompleteness in requirements. [Principal ideas/results] We mask words in requirements and have BERT's masked language model (MLM) generate contextualized predictions for filling the masked slots. We simulate incompleteness by withholding content from requirements and measure BERT's ability to predict terminology that is present in the withheld content but absent in the content disclosed to BERT. [Contribution] BERT can be configured to generate multiple predictions per mask. Our first contribution is to determine how many predictions per mask is an optimal trade-off between effectively discovering omissions in requirements and the level of noise in the predictions. Our second contribution is devising a machine learning-based filter that post-processes predictions made by BERT to further reduce noise. We empirically evaluate our solution over 40 requirements specifications drawn from the PURE dataset [1]. Our results indicate that: (1) predictions made by BERT are highly effective at pinpointing terminology that is missing from requirements, and (2) our filter can substantially reduce noise from the predictions, thus making BERT a more compelling aid for improving completeness in requirements.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/03/2023

Improving Requirements Completeness: Automated Assistance through Large Language Models

Natural language (NL) is arguably the most prevalent medium for expressi...
research
02/09/2023

AI-based Question Answering Assistance for Analyzing Natural-language Requirements

By virtue of being prevalently written in natural language (NL), require...
research
08/23/2023

Bridging the Gap: Deciphering Tabular Data Using Large Language Model

In the realm of natural language processing, the understanding of tabula...
research
05/13/2021

Are Larger Pretrained Language Models Uniformly Better? Comparing Performance at the Instance Level

Larger language models have higher accuracy on average, but are they bet...
research
05/10/2020

How Context Affects Language Models' Factual Predictions

When pre-trained on large unsupervised textual corpora, language models ...
research
08/24/2023

Requirements Quality Assurance in Industry: Why, What and How?

Context and Motivation: Natural language is the most common form to spec...

Please sign up or login with your details

Forgot password? Click here to reset