Poster (Invited) FZJ-2025-05415

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
LeapFactual: Reliable Visual Counterfactual Explanation Using Conditional Flow Matching

 ;  ;  ;  ;

2025

The Thirty-Ninth Annual Conference on Neural Information Processing, NeurIPS2025, San DiegoSan Diego, USA, 1 Dec 2025 - 7 Dec 20252025-12-012025-12-07 [10.34734/FZJ-2025-05415]

This record in other databases:  

Please use a persistent id in citations: doi:

Abstract: The growing integration of machine learning (ML) and artificial intelligence (AI) models into high-stakes domains such as healthcare and scientific research calls for models that are not only accurate but also interpretable. Among the existing explainable methods, counterfactual explanations offer interpretability by identifying minimal changes to inputs that would alter a model's prediction, thus providing deeper insights. However, current counterfactual generation methods suffer from critical limitations, including gradient vanishing, discontinuous latent spaces, and an overreliance on the alignment between learned and true decision boundaries. To overcome these limitations, we propose LeapFactual, a novel counterfactual explanation algorithm based on conditional flow matching. LeapFactual generates reliable and informative counterfactuals, even when true and learned decision boundaries diverge. Following a model-agnostic approach, LeapFactual is not limited to models with differentiable loss functions. It can even handle human-in-the-loop systems, expanding the scope of counterfactual explanations to domains that require the participation of human annotators, such as citizen science. We provide extensive experiments on benchmark and real-world datasets showing that LeapFactual generates accurate and in-distribution counterfactual explanations that offer actionable insights. We observe, for instance, that our reliable counterfactual samples with labels aligning to ground truth can be beneficially used as new training data to enhance the model. The proposed method is broadly applicable and enhances both scientific knowledge discovery and non-expert interpretability.


Contributing Institute(s):
  1. Datenanalyse und Maschinenlernen (IAS-8)
Research Program(s):
  1. 5254 - Neuroscientific Data Analytics and AI (POF4-525) (POF4-525)
  2. 5112 - Cross-Domain Algorithms, Tools, Methods Labs (ATMLs) and Research Groups (POF4-511) (POF4-511)

Appears in the scientific report 2025
Database coverage:
OpenAccess
Click to display QR Code for this record

The record appears in these collections:
Document types > Presentations > Poster
Institute Collections > IAS > IAS-8
Workflow collections > Public records
Publications database
Open Access

 Record created 2025-12-16, last modified 2025-12-17


OpenAccess:
Download fulltext PDF
Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)