Abstract
[Context and Motivation] Domain models are valuable tools for facilitating software development by providing structured insights into the problem space. [Question/problem] Building domain models requires time, domain, and modeling expertise; therefore, it is an expensive task. Due to this cost, domain models are often overlooked in agile development, where the primary form of documentation is user stories that capture actors, their desires, and, optionally, the rationale. [Principal ideas] Automated domain model generation from user stories can support the software development process in agile settings. Recent advances in large language models offer promise in automating such tasks through improved language understanding and reasoning. This research preview investigates the potential of open-source large language models to automatically generate domain models from natural language requirements, employing instruction tuning on a substantial dataset of user stories and their corresponding domain models. By comparing instruction-tuned and pre-trained models, we aim to assess the specific impact of instruction tuning on model generation quality. [Contributions] This study is, to our knowledge, the first to apply instruction tuning on large language models specifically for domain model generation. Through qualitative and quantitative evaluations of factors like completeness and correctness, we assess the performance and quality of the instruction-tuned models.
Original language | English |
---|---|
Title of host publication | Requirements Engineering: Foundation for Software Quality |
Subtitle of host publication | 31st International Working Conference, REFSQ 2025, Barcelona, Spain, April 7–10, 2025, Proceedings |
Editors | Anne Hess, Angelo Susi |
Publisher | Springer |
Pages | 157-165 |
Number of pages | 9 |
ISBN (Electronic) | 978-3-031-88531-0 |
ISBN (Print) | 978-3-031-88530-3 |
DOIs | |
Publication status | Published - 1 Apr 2025 |
Publication series
Name | Lecture Notes in Computer Science |
---|---|
Volume | 15588 |
ISSN (Print) | 0302-9743 |
ISSN (Electronic) | 1611-3349 |
Bibliographical note
Publisher Copyright:© The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
Keywords
- Domain models
- Generative AI
- Instruction Tuning
- Large language models