TY - GEN
T1 - Leveraging ChatGPT to Predict Requirements Testability with Differential In-Context Learning
AU - Dahiya, Mahima
AU - Gill, Rashminder
AU - Niu, Nan
AU - Gudaparthi, Hemanth
AU - Peng, Zedong
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Testability is a desired property of requirements, indicating how easy or difficult a requirements artifact supports its own testing. Prior work predicts natural language (NL) requirements' testability by training a decision tree (DT) via some readability and word measures. To explore better ways of predicting requirements testability, we examine in this paper large language models-ChatGPT in particular. Our experiments on a total of 1,181 requirements from six software systems show that ChatGPT's zero-shot learning performs worse than the DT. A main reason is due to the lack of context specific to the testability prediction task. However, applying ChatGPT's incontext learning (ICL) reveals a limitation of skewed examples caused by the imbalanced data. Thus, we propose a novel approach, called differential ICL, to address the challenges by exploiting the DT and show quantitatively the higher accuracy achieved by differential ICL.
AB - Testability is a desired property of requirements, indicating how easy or difficult a requirements artifact supports its own testing. Prior work predicts natural language (NL) requirements' testability by training a decision tree (DT) via some readability and word measures. To explore better ways of predicting requirements testability, we examine in this paper large language models-ChatGPT in particular. Our experiments on a total of 1,181 requirements from six software systems show that ChatGPT's zero-shot learning performs worse than the DT. A main reason is due to the lack of context specific to the testability prediction task. However, applying ChatGPT's incontext learning (ICL) reveals a limitation of skewed examples caused by the imbalanced data. Thus, we propose a novel approach, called differential ICL, to address the challenges by exploiting the DT and show quantitatively the higher accuracy achieved by differential ICL.
KW - large language models
KW - machine learning
KW - requirements testability
KW - software testing
UR - https://www.scopus.com/pages/publications/85207850923
U2 - 10.1109/IRI62200.2024.00044
DO - 10.1109/IRI62200.2024.00044
M3 - Conference contribution
AN - SCOPUS:85207850923
T3 - Proceedings - 2024 IEEE International Conference on Information Reuse and Integration for Data Science, IRI 2024
SP - 170
EP - 175
BT - Proceedings - 2024 IEEE International Conference on Information Reuse and Integration for Data Science, IRI 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 25th IEEE International Conference on Information Reuse and Integration for Data Science, IRI 2024
Y2 - 7 August 2024 through 9 August 2024
ER -