Abstract
Recent advancements in large language models (LLMs) have shown promise in generating psychotherapeutic dialogues, particularly in the context of motivational interviewing (MI). However, the inherent lack of transparency in LLM outputs presents significant challenges given the sensitive nature of psychotherapy. Applying MI strategies, a set of MI skills, to generate more controllable therapeutic-adherent conversations with explainability provides a possible solution. In this work, we explore the alignment of LLMs with MI strategies by first prompting the LLMs to predict the appropriate strategies as reasoning and then utilizing these strategies to guide the subsequent dialogue generation. We seek to investigate whether such alignment leads to more controllable and explainable generations. Multiple experiments including automatic and human evaluations are conducted to validate the effectiveness of MI strategies in aligning psychotherapy dialogue generation. Our findings demonstrate the potential of LLMs in producing strategically aligned dialogues and suggest directions for practical applications in psychotherapeutic settings.
Original language | English |
---|---|
Title of host publication | Main Conference |
Editors | Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 1983-2002 |
Number of pages | 20 |
ISBN (Electronic) | 9798891761964 |
Publication status | Published - 2025 |
Event | 31st International Conference on Computational Linguistics, COLING 2025 - Abu Dhabi, United Arab Emirates Duration: 19 Jan 2025 → 24 Jan 2025 |
Publication series
Name | Proceedings - International Conference on Computational Linguistics, COLING |
---|---|
Volume | Part F206484-1 |
ISSN (Print) | 2951-2093 |
Conference
Conference | 31st International Conference on Computational Linguistics, COLING 2025 |
---|---|
Country/Territory | United Arab Emirates |
City | Abu Dhabi |
Period | 19/01/25 → 24/01/25 |
Bibliographical note
Publisher Copyright:© 2025 Association for Computational Linguistics.