Authors:
Reine Santos
1
;
Gabriel Freitas
1
;
Igor Steinmacher
2
;
Tayana Conte
1
;
Ana Oran
1
and
Bruno Gadelha
1
Affiliations:
1
Instituto de Computação (ICOMP), Universidade Federal do Amazonas (UFAM), Av. Gal. Rodrigo Octávio Jordão Ramos, Manaus, Brazil
;
2
Department of Computer Science, Northern Arizona University (NAU), 1900 S Knoles Dr, Flagstaff, AZ 86011, U.S.A.
Keyword(s):
User Story, Large Language Models, Requirements Engineering, Information System.
Abstract:
In agile software development, user stories play a central role in defining system requirements, fostering communication, and guiding development efforts. Despite their importance, they are often poorly written, exhibiting quality defects that hinder project outcomes and reduce team efficiency. Manual methods for creating user stories are time-consuming and prone to errors and inconsistencies. Advancements in Large Language Models (LLMs), such as ChatGPT, present a promising avenue for automating and improving this process. This research explores whether user stories generated by ChatGPT, using prompting techniques, achieve higher quality than those created manually by humans. User stories were assessed using the Quality User Story (QUS) framework. We conducted two empirical studies to address this. The first study compared manually created user stories with those generated by ChatGPT through free-form prompt. This study involved 30 participants and found no statistically significant
difference between the two methods. The second study compared free-form prompt with meta-few-shot prompt, demonstrating that the latter outperformed both, achieving higher consistency and semantic quality with an efficiency calculated based on the success rate of 88.57%. These findings highlight the potential of LLMs with prompting techniques to enhance user story generation, offering a reliable and effective alternative to traditional methods.
(More)