Results
- Showing results for:
- Reset all filters
Search results
-
Conference paperGorur D, Rago A, Toni F, 2025,
Argumentatively coherent judgmental forecasting
, 28th European Conference on Artificial Intelligence, Publisher: IOS PressJudgmental forecasting employs human opinions to make predictions about future events, rather than exclusively historical data as in quantitative forecasting. When these opinions form an argumentative structure around forecasts, it is useful to study the properties of the forecasts from an argumentative perspective. In this paper, we advocate and formally define a property of argumentative coherence, which, in essence, requires that a forecaster’s reasoning iscoherent with their forecast. We then conduct three evaluations with our notion of coherence. First, we assess the impact of enforcing coherence on human forecasters as well as on Large Language Model (LLM)-based forecasters, given that they have recently shown to be competitive with human forecasters. In both cases, we show that filtering out incoherent predictions improves forecasting accuracy consistently, supporting the practical value of coherence in both human and LLM-based forecasting. Then, via crowd-sourced user experiments, we show that, despite its apparent intuitiveness and usefulness, users do not generally align with this coherence property. This points to the need to integrate, within argumentation-based judgmental forecasting, mechanisms to filter out incoherent opinions before obtaining group forecasting predictions.
-
Conference paperGigante N, Leofante F, Micheli A, 2025,
Counterfactual scenarios for automated planning
, 22nd International Conference on Principles of Knowledge Representation and Reasoning, Publisher: International Joint Conferences on Artificial Intelligence OrganizationCounterfactual Explanations (CEs) are a powerful technique used to explain Machine Learning models by showing how the input to a model should be minimally changed for the model to produce a different output. Similar proposals havebeen made in the context of Automated Planning, where CEs have been characterised in terms of minimal modifications to an existing plan that would result in the satisfaction of a different goal. While such explanations may help diagnose faults and reason about the characteristics of a plan, they fail to capture higher-level properties of the problem being solved. To address this limitation, we propose a novel explanation paradigm that is based on counterfactual scenarios. In particular, given a planning problem P and an LTLf formula ψ defining desired properties of a plan, counterfactual scenarios identify minimal modifications to P such that it admits plans that comply with ψ. In this paper, we present two qualitative instantiations of counterfactual scenarios based on an explicit quantification over plans that must satisfy ψ. We then characterise the computational complexity of generating such counterfactual scenarios when different types of changes are allowed on P. We show that producing counterfactual scenarios is often only as expensive as computing a plan for P , thus demonstrating the practical viability of our proposal andultimately providing a framework to construct practical algorithms in this area.
-
Conference paperRago A, Palfi B, Sukpanichnant P, et al., 2025,
Exploring the effect of explanation content and format on user comprehension and trust in healthcare
, The 14th Conference on Prestigious Applications of Intelligent Systems (PAIS 2025), Publisher: IOS PressAI-driven tools for healthcare are widely acknowledgedas potentially beneficial to health practitioners and patients, e.g. the QCancer regression tool for cancer risk prediction. However, for these tools to be trusted, they need to be supplemented with explanations. We examine how explanations’ content and format affect user comprehension and trust when explaining QCancer’s predictions. Regarding content, we deploy SHAP and Occlusion-1. Regarding format, we present SHAP explanations, conventionally, ascharts (SC) and Occlusion-1 explanations as charts (OC) as well as text (OT), to which their simpler nature lends itself. We conduct experiments with two sets of stakeholders: the general public (representing patients) and medical students (representing healthcare practitioners). Our experiments showed higher subjective comprehension and trust for Occlusion-1 over SHAP explanations based on content.However, when controlling for format, only OT outperformed SC, suggesting this trend is driven by preferences for text. Other findings corroborated that explanation format, rather than content, is often the critical factor.
-
Conference paperRapberger A, Russo F, Rago A, et al., 2025,
On gradual semantics for assumption-based argumentation
, 22nd International Conference on Principles of Knowledge Representation and Reasoning (KR 2025)In computational argumentation, gradual semantics are fine-grained alternatives to extension-based and labelling-basedsemantics. They ascribe a dialectical strength to (components of) arguments sanctioning their degree of acceptability. Several gradual semantics have been studied for abstract, bipolar and quantitative bipolar argumentation frameworks (QBAFs), as well as, to a lesser extent, for some forms of structured argumentation. However, this has not been the case for assumption-based argumentation (ABA), despite it being a popular form of structured argumentation with several applications where gradual semantics could be useful. In this paper, we fill this gap and propose a family of novel gradual semantics for equipping assumptions, which are the core components in ABA frameworks, with dialectical strengths. To do so, we use bipolar set-based argumentation frame-works as an abstraction of (potentially non-flat) ABA frame-works and generalise state-of-the-art modular gradual semantics for QBAFs. We show that our gradual ABA semantics satisfy suitable adaptations of desirable properties of gradual QBAF semantics, such as balance and monotonicity. We also explore an argument-based approach that leverages established QBAF modular semantics directly, and use it as base-line. Finally, we conduct experiments with synthetic ABA frameworks to compare our gradual ABA semantics with its argument-based counterpart and assess convergence.
-
Conference paperRaaghav V, Bikos D, Rago A, et al., 2025,
Explainable prediction of the mechanical properties of composites with CNNs
, The 14th Conference on Prestigious Applications of Intelligent Systems (PAIS 2025), Publisher: IOS PressComposites are amongst the most important materialsmanufactured today, as evidenced by their use in countless applications. In order to establish the suitability of composites in specific applications, finite element (FE) modelling, a numerical method based on partial differential equations, is the industry standard for assessing their mechanical properties. However, FE modelling is exceptionally costly from a computational viewpoint, a limitation which has led to efforts towards applying AI models to this task. However, in these approaches: the chosen model architectures were rudimentary, feed-forward neural networks giving limited accuracy; the studies focus on predicting elastic mechanical properties, without considering material strength limits; and the models lacked transparency, hindering trustworthiness by users. In this paper, we show that convolutional neural networks (CNNs) equipped with methods from explainable AI (XAI) can be successfully deployed to solve this problem. Ourapproach uses customised CNNs trained on a dataset we generate using transverse tension tests in FE modelling to predict composites’ mechanical properties, i.e., Young’s modulus and yield strength. We show empirically that our approach achieves high accuracy, outperforming a baseline, ResNet-34, in estimating the mechanical properties. We then use SHAP and Integrated Gradients, two post-hocXAI methods, to explain the predictions, showing that the CNNs use the critical geometrical features that influence the composites’ behaviour, thus allowing engineers to verify that the models are trust-worthy by representing the science of composites.
-
Conference paperRago A, Vasileiou SL, Tran S, et al., 2025,
A methodology for incompleteness-tolerant and modular gradual semantics for argumentative statement graphs
, 22nd International Conference on Principles of Knowledge Representation and Reasoning (KR 2025), Publisher: International Joint Conferences on Artificial Intelligence OrganizationGradual semantics (GS) have demonstrated great potential in argumentation, in particular for deploying quantitative bipolar argumentation frameworks (QBAFs) in a number of real-world settings, from judgmental forecasting to explainable AI. In this paper, we provide a novel methodology for obtaining GS for statement graphs, a form of structured argumentation framework, where arguments and relations between them are built from logical statements. Our methodology differs from existing approaches in the literature in two main ways. First, it naturally accommodates incomplete information, so that arguments with partially specified premises can play ameaningful role in the evaluation. Second, it is modularlydefined to leverage on any GS for QBAFs. We also define aset of novel properties for our GS and study their suitabilityalongside a set of existing properties (adapted to our setting) for two instantiations of our GS, demonstrating their advantages over existing approaches.
-
Journal articleDickie C, Lauren S, Belardinelli F, et al., 2025,
Aggregating bipolar opinions through bipolar assumption-based argumentation
, Autonomous Agents and Multi-Agent Systems, Vol: 39, ISSN: 1387-2532We introduce a novel method to aggregate Bipolar ArgumentationFrameworks expressing opinions of different parties in debates. We use BipolarAssumption-based Argumentation (ABA) as an all-encompassing formalismfor Bipolar Argumentation under different semantics. By leveraging on recentresults on judgement aggregation in Social Choice Theory, we prove severalpreservation results for relevant properties of Bipolar ABA using quota andoligarchic rules. Specifically, we prove (positive and negative) results about thepreservation of conflict-free, closed, admissible, preferred, complete, set-stable,well-founded and ideal extensions in Bipolar ABA, as well as the preservationof acceptability, acyclicity and coherence for individual assumptions. Finally,we illustrate our methodology and results in the context of a case study onopinion aggregation for the treatment of long COVID patients.
-
Conference paperFreedman G, Toni F, 2025,
Exploring the potential for large language models to demonstrate rational probabilistic beliefs
, 38th International FLAIRS Conference, Publisher: LibraryPress@UF, ISSN: 2334-0762Advances in the general capabilities of large language models (LLMs) have led to their use for information retrieval, and as components in automated decision systems. A faithful representation of probabilistic reasoning in these models may be essential to ensure trustworthy, explainable and effective performance in these tasks. Despite previous work suggesting that LLMs can perform complex reasoning and well-calibrated uncertainty quantification, we find that current versions of this class of model lack the ability to provide rational and coherent representations of probabilistic beliefs. To demonstrate this, we introduce a novel dataset of claims with indeterminate truth values and apply a number of well-established techniques for uncertainty quantification to measure the ability of LLM's to adhere to fundamental properties of probabilistic reasoning.
-
Conference paperAlfano G, Gould A, Leofante F, et al., 2025,
Counterfactual explanations under model multiplicity and their use in computational argumentation
, International Joint Conference on Artificial Intelligence (IJCAI) 2025, Publisher: IJCAICounterfactual explanations (CXs) are widely recognised as an essential technique for providing recourse recommendations for AI models. However, it is not obvious how to determine CXs in model multiplicity scenarios, where equally performing but different models can be obtained forthe same task. In this paper, we propose novel qualitative and quantitative definitions of CXs based on explicit, nested quantification over (groups) of model decisions. We also study properties of these notions and identify decision problems of interest therefor. While our CXs are broadly applicable, in this paper we instantiate them within computational argumentation where model multiplicity naturally emerges e.g. with incomplete and case-based argumentation frameworks. We then illustrate the suitability of our CXs for model multiplicity in legal and healthcare contexts, before analysing the complexity of the associated decision problems.
-
Conference paperAyoobi H, Potyka N, Toni F, 2025,
ProtoArgNet: Interpretable Image Classification with Super-Prototypes and Argumentation
, AAAI Conference on Artificial Intelligence -
Conference paperChen L, Dejl A, Toni F, 2025,
Identifying Query-Relevant Neurons in Large Language Models for Long-Form Texts
, The 39th Annual AAAI Conference on Artificial Intelligence -
Conference paperFreedman G, Dejl A, Gorur D, et al., 2025,
Argumentative large language models for explainable and contestable claim verification
, AAAI Conference on Artificial Intelligence, Publisher: Association for the Advancement of Artificial Intelligence, Pages: 14930-14939, ISSN: 2159-5399The profusion of knowledge encoded in large language models (LLMs) and their ability to apply this knowledge zero-shot in a range of settings makes them promising candidates for use in decision-making. However, they are currently limited by their inability to provide outputs which can be faithfully explained and effectively contested to correct mistakes. In this paper, we attempt to reconcile these strengths and weaknesses by introducing argumentative LLMs (ArgLLMs), a method for augmenting LLMs with argumentative reasoning. Concretely, ArgLLMs construct argumentation frameworks, which then serve as the basis for formal reasoning in support of decision-making. The interpretable nature of these argumentation frameworks and formal reasoning means that any decision made by ArgLLMs may be explained and contested. We evaluate ArgLLMs’ performance experimentally in comparison with state-of-the-art techniques, in the context of the decision-making task of claim verification. We also define novel properties to characterise contestability and assess ArgLLMs formally in terms of these properties.
-
Conference paperRusso F, Toni F, 2025,
Shapley-PC: constraint-based causal structure learning with a Shapley inspired framework
, 4th Conference on Causal Learning and Reasoning (CLeaR 2025)Causal Structure Learning (CSL), also referred to as causal discovery, amounts to extracting causal relations among variables in data. CSL enables the estimation of causal effects from observational data alone, avoiding the need to perform real life experiments. Constraint-based CSL leverages conditional independence tests to perform causal discovery. We propose Shapley-PC, a novel method to improve constraint-based CSL algorithms by using Shapley values over the possible conditioning sets, to decide which variables are responsible for the observed conditional (in)dependences. We prove soundness, completeness and asymptotic consistency of Shapley-PC and run a simulationstudy showing that our proposed algorithm is superior to existing versions of PC.
-
Conference paperKori A, Glocker B, Toni F, 2025,
Explaining Image Classifiers with Visual Debates
, 27th International Conference on Discovery Science, Publisher: SPRINGER INTERNATIONAL PUBLISHING AG, Pages: 200-214, ISSN: 2945-9133- Cite
- Citations: 1
-
Conference paperDe Angelis E, Proietti M, Toni F, 2025,
Greedy ABA Learning for Case-Based Reasoning
, 24th International Conference on Autonomous Agents and Multiagent Systems-AAMAS-Annual, Publisher: ASSOC COMPUTING MACHINERY, Pages: 556-564 -
Conference paperKori A, Rago A, Toni F, 2025,
Free argumentative exchanges for explaining image classifiers
, AAMAS 2025, Publisher: ACMDeep learning models are powerful image classifiers but their opacity hinders their trustworthiness. Explanation methods for capturing the reasoning process within these classifiers faithfully and in a cognitively manageable manner are scarce, due to their sheer complexity and size. In this paper, we provide a solution for this problem by defining a novel method for explaining the outputs of image classifiers with debates between two agents, each arguing for a particular class. We obtain these debates as concrete instances of Free Argumentative eXchanges (FAXs), a novel argumentation-based multi-agent framework allowing agents to internalise opinions by other agents differently than originally stated. We define two metrics to assess the usefulness of FAXs as argumentative explanationsfor image classifiers. We then conduct a number of empirical experiments showing that FAXs perform well along these metrics as well as being more faithful to the image classifiers than conventional, non-argumentative explanation methods.
-
Conference paperGorur D, Rago A, Toni F, 2025,
Can Large Language Models perform Relation-based Argument Mining?
, The 31st International Conference on Computational Linguistics -
Conference paperRapberger A, Ulbricht M, Toni F, 2024,
On the correspondence of non-flat assumption-based argumentation and logic programming with negation as failure in the head
, 22nd International Workshop on Nonmonotonic Reasoning NMR 24), Publisher: CEUR Workshop Proceedings, Pages: 122-121, ISSN: 1613-0073The relation between (a fragment of) assumption-based argumentation (ABA) and logic programs (LPs) under stable model semantics is well-studied. However, for obtaining this relation, the ABA framework needs to be restricted to being flat, i.e., a fragment where the (defeasible) assumptions can never be entailed, only assumed to be true or false. Here, we remove this restriction and show a correspondence between non-flat ABA and LPs with negation as failure in their head. We then extend this result to so-called set-stable ABA semantics, originally defined for the fragment of non-flat ABA called bipolar ABA. We showcase how to define set-stable semantics for LPs with negation as failure in their head and show the correspondence to set-stable ABA semantics.
-
Conference paperVasileiou S, Kumar A, Yeoh W, et al., 2024,
Dialectical reconciliation via structured argumentative dialogues
, KR 2024We present a novel framework designed to extend model reconciliation approaches, commonly used in human-aware planning, for enhanced human-AI interaction. By adopting a structured argumentation-based dialogue paradigm, our framework enables dialectical reconciliation to address knowledge discrepancies between an explainer (AI agent) and an explainee (human user), where the goal is for the explainee to understand the explainer's decision. We formally describe the operational semantics of our proposed framework, providing theoretical guarantees. We then evaluate the framework's efficacy ``in the wild'' via computational and human-subject experiments. Our findings suggest that our framework offers a promising direction for fostering effective human-AI interactions in domains where explainability is important.
-
Conference paperBattaglia E, Baroni P, Rago A, et al., 2024,
Integrating user preferences into gradual bipolar argumentation for personalised decision support
, Scalable Uncertainty Management, 16th International Conference (SUM 2024), Publisher: Springer, Pages: 14-28, ISSN: 1611-3349Gradual bipolar argumentation has been shown to be aneffective means for supporting decisions across a number of domains. Individual user preferences can be integrated into the domain knowledge represented by such argumentation frameworks and should be taken into account in order to provide personalised decision support. This howeverrequires the definition of a suitable method to handle user-provided preferences in gradual bipolar argumentation, which has not been considered in previous literature. Towards filling this gap, we develop a conceptual analysis on the role of preferences in argumentation and investigate some basic principles concerning the effects they should have on the evaluation of strength in gradual argumentation semantics. We illustrate an application of our approach in the context of a review aggregation system, which has been enhanced with the ability to produce personalisedoutcomes based on user preferences.
This data is extracted from the Web of Science and reproduced under a licence from Thomson Reuters. You may not copy or re-distribute this data in whole or in part without the written consent of the Science business of Thomson Reuters.
Contact us
Artificial Intelligence Network
South Kensington Campus
Imperial College London
SW7 2AZ
To reach the elected speaker of the network, Dr Rossella Arcucci, please contact:
To reach the network manager, Diana O'Malley - including to join the network - please contact: