TY - GEN
T1 - Application of back-translation
T2 - 2021 ACM Southeast Conference, ACMSE 2021
AU - Subedi, Ishan Mani
AU - Singh, Maninder
AU - Ramasamy, Vijayalakshmi
AU - Walia, Gursimran Singh
N1 - Publisher Copyright:
© 2021 ACM.
PY - 2021/4/15
Y1 - 2021/4/15
N2 - Ambiguous requirements are problematic in requirement engineering as various stakeholders can debate on the interpretation of the requirements leading to a variety of issues in the development stages. Since requirement specifications are usually written in natural language, analyzing ambiguous requirements is currently a manual process as it has not been fully automated to meet the industry standards. In this paper, we used transfer learning by using ULMFiT where we pre-trained our model to a general-domain corpus and then fine-tuned it to classify ambiguous vs unambiguous requirements (target task). We then compared its accuracy with machine learning classifiers like SVM, Linear Regression, and Multinomial Naive Bayes. We also used back translation (BT) as a text augmentation technique to see if it improved the classification accuracy. Our results showed that ULMFiT achieved higher accuracy than SVM (Support Vector Machines), Logistic Regression and Multinomial Naive Bayes for our initial data set. Further by augmenting requirements using BT, ULMFiT got a higher accuracy than SVM, Logistic Regression, and Multinomial Naive Bayes classifier, improving the initial performance by 5.371%. Our proposed research provides some promising insights on how transfer learning and text augmentation can be applied to small data sets in requirements engineering.
AB - Ambiguous requirements are problematic in requirement engineering as various stakeholders can debate on the interpretation of the requirements leading to a variety of issues in the development stages. Since requirement specifications are usually written in natural language, analyzing ambiguous requirements is currently a manual process as it has not been fully automated to meet the industry standards. In this paper, we used transfer learning by using ULMFiT where we pre-trained our model to a general-domain corpus and then fine-tuned it to classify ambiguous vs unambiguous requirements (target task). We then compared its accuracy with machine learning classifiers like SVM, Linear Regression, and Multinomial Naive Bayes. We also used back translation (BT) as a text augmentation technique to see if it improved the classification accuracy. Our results showed that ULMFiT achieved higher accuracy than SVM (Support Vector Machines), Logistic Regression and Multinomial Naive Bayes for our initial data set. Further by augmenting requirements using BT, ULMFiT got a higher accuracy than SVM, Logistic Regression, and Multinomial Naive Bayes classifier, improving the initial performance by 5.371%. Our proposed research provides some promising insights on how transfer learning and text augmentation can be applied to small data sets in requirements engineering.
KW - Machine learning
KW - Neural networks
KW - Requirement engineering and quality
KW - Transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85106428321&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85106428321&partnerID=8YFLogxK
U2 - 10.1145/3409334.3452068
DO - 10.1145/3409334.3452068
M3 - Conference contribution
AN - SCOPUS:85106428321
T3 - Proceedings of the 2021 ACMSE Conference - ACMSE 2021: The Annual ACM Southeast Conference
SP - 130
EP - 137
BT - Proceedings of the 2021 ACMSE Conference - ACMSE 2021
PB - Association for Computing Machinery, Inc
Y2 - 15 April 2021 through 17 April 2021
ER -