Generating Domain Models with LLMs Using Instruction Tuning: A Research Preview

Gökberk Çelikmasat*, Atay Özgövde, Fatma Başak Aydemir

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

[Context and Motivation] Domain models are valuable tools for facilitating software development by providing structured insights into the problem space. [Question/problem] Building domain models requires time, domain, and modeling expertise; therefore, it is an expensive task. Due to this cost, domain models are often overlooked in agile development, where the primary form of documentation is user stories that capture actors, their desires, and, optionally, the rationale. [Principal ideas] Automated domain model generation from user stories can support the software development process in agile settings. Recent advances in large language models offer promise in automating such tasks through improved language understanding and reasoning. This research preview investigates the potential of open-source large language models to automatically generate domain models from natural language requirements, employing instruction tuning on a substantial dataset of user stories and their corresponding domain models. By comparing instruction-tuned and pre-trained models, we aim to assess the specific impact of instruction tuning on model generation quality. [Contributions] This study is, to our knowledge, the first to apply instruction tuning on large language models specifically for domain model generation. Through qualitative and quantitative evaluations of factors like completeness and correctness, we assess the performance and quality of the instruction-tuned models.

Original languageEnglish
Title of host publicationRequirements Engineering: Foundation for Software Quality
Subtitle of host publication31st International Working Conference, REFSQ 2025, Barcelona, Spain, April 7–10, 2025, Proceedings
EditorsAnne Hess, Angelo Susi
PublisherSpringer
Pages157-165
Number of pages9
ISBN (Electronic)978-3-031-88531-0
ISBN (Print)978-3-031-88530-3
DOIs
Publication statusPublished - 1 Apr 2025

Publication series

NameLecture Notes in Computer Science
Volume15588
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Bibliographical note

Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.

Keywords

  • Domain models
  • Generative AI
  • Instruction Tuning
  • Large language models

Fingerprint

Dive into the research topics of 'Generating Domain Models with LLMs Using Instruction Tuning: A Research Preview'. Together they form a unique fingerprint.

Cite this