Parameterized Argumentation-based Reasoning Tasks for Benchmarking Generative Language Models

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

Generative large language models as tools in the legal domain have the potential to improve the justice system. However, the reasoning behavior of current generative models is brittle and poorly understood, hence cannot be responsibly applied in the domains of law and evidence. This paper presents reasoning benchmarks that are dynamically varied, scalable in their complexity, and have formally unambiguous interpretations. In this study, we illustrate the approach on the basis of witness testimony, focusing on the underlying argument attack structure. We dynamically generate both linear and non-linear argument attack graphs of varying complexity and translate these into reasoning puzzles about witness testimony expressed in natural language. We show that state-of-the-art large language models often fail in these reasoning puzzles, already at low complexity. Obvious mistakes are made by the models, and their inconsistent performance indicates that their reasoning capabilities are brittle. Furthermore, at higher complexity, even state-of-the-art models specifically designed for reasoning make mistakes. We show the viability of using a parametrized benchmark with varying complexity to evaluate the reasoning capabilities of generative language models, which contribute to a better understanding of the limitations of the reasoning capabilities of generative models.
Original languageEnglish
Title of host publicationProceedings of the Twentieth International Conference on Artificial Intelligence and Law
EditorsJuliano Maranhão
PublisherAssociation for Computing Machinery
Pages455-459
Number of pages5
ISBN (Print)979-8-4007-1939-4
DOIs
Publication statusPublished - 13 Jan 2026
EventInternational Conference on Artificial Intelligence and Law - Chicago, United States
Duration: 16 Jun 202520 Jun 2025
Conference number: 20

Conference

ConferenceInternational Conference on Artificial Intelligence and Law
Abbreviated titleICAIL
Country/TerritoryUnited States
CityChicago
Period16/06/2520/06/25

Bibliographical note

Publisher Copyright:
© 2025 Copyright held by the owner/author(s).

Keywords

  • LLMs
  • argumentation
  • benchmarks
  • generative AI
  • reasoning

Fingerprint

Dive into the research topics of 'Parameterized Argumentation-based Reasoning Tasks for Benchmarking Generative Language Models'. Together they form a unique fingerprint.

Cite this