TY - GEN
T1 - Deriving Domain Models From User Stories
T2 - 32nd IEEE International Requirements Engineering Conference, RE 2024
AU - Bragilovski, Maxim
AU - Van Can, Ashley T.
AU - Dalpiaz, Fabiano
AU - Sturm, Arnon
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/8/21
Y1 - 2024/8/21
N2 - Domain models play a crucial role in software development, as they provide means for communication among stakeholders, for eliciting requirements, and for representing the information structure behind a database scheme or at the basis of model-driven development. However, creating such models is a tedious activity and automated support may assist in obtaining an initial domain model that can later be enriched by human analysts. In this paper, we propose an experimental comparison of the effectiveness of various approaches for deriving domain models from a given set of user stories. We contrast human derivation with machine derivation; for the latter, we compare (i) the Visual Narrator: an existing rule-based NLP approach; (ii) a machine-learning classifier that we feature engineered; and (iii) a generative AI approach that we constructed via prompt engineering. Based on a benchmark dataset that consists of nine collections of user stories and corresponding domain models, the evaluation indicates that no approach matches human performance, although a tuned version of the machine learning approach comes close. To better understand the results, we qualitatively analyze them and identify differences in the types of false positives as well as other factors that affect performance.
AB - Domain models play a crucial role in software development, as they provide means for communication among stakeholders, for eliciting requirements, and for representing the information structure behind a database scheme or at the basis of model-driven development. However, creating such models is a tedious activity and automated support may assist in obtaining an initial domain model that can later be enriched by human analysts. In this paper, we propose an experimental comparison of the effectiveness of various approaches for deriving domain models from a given set of user stories. We contrast human derivation with machine derivation; for the latter, we compare (i) the Visual Narrator: an existing rule-based NLP approach; (ii) a machine-learning classifier that we feature engineered; and (iii) a generative AI approach that we constructed via prompt engineering. Based on a benchmark dataset that consists of nine collections of user stories and corresponding domain models, the evaluation indicates that no approach matches human performance, although a tuned version of the machine learning approach comes close. To better understand the results, we qualitatively analyze them and identify differences in the types of false positives as well as other factors that affect performance.
KW - Domain Models
KW - Large Language Models
KW - Machine Learning
KW - Model Derivation
KW - Requirements Engineering
KW - User Stories
UR - http://www.scopus.com/inward/record.url?scp=85202714739&partnerID=8YFLogxK
U2 - 10.1109/RE59067.2024.00014
DO - 10.1109/RE59067.2024.00014
M3 - Conference contribution
AN - SCOPUS:85202714739
T3 - Proceedings of the IEEE International Conference on Requirements Engineering
SP - 31
EP - 42
BT - Proceedings - 32nd IEEE International Requirements Engineering Conference, RE 2024
A2 - Liebel, Grischa
A2 - Hadar, Irit
A2 - Spoletini, Paola
PB - IEEE Computer Society
Y2 - 24 June 2024 through 28 June 2024
ER -