Translating theory into practice: assessing the privacy implications of concept-based explanations for biomedical AI.

Front Bioinform

Smart Data and Knowledge Services (SDS), Deutsches Forschungszentrum für Künstliche Intelligenz (DFKI) GmbH, Kaiserslautern, Germany.

Published: July 2023

Artificial Intelligence (AI) has achieved remarkable success in image generation, image analysis, and language modeling, making data-driven techniques increasingly relevant in practical real-world applications, promising enhanced creativity and efficiency for human users. However, the deployment of AI in high-stakes domains such as infrastructure and healthcare still raises concerns regarding algorithm accountability and safety. The emerging field of explainable AI (XAI) has made significant strides in developing interfaces that enable humans to comprehend the decisions made by data-driven models. Among these approaches, concept-based explainability stands out due to its ability to align explanations with high-level concepts familiar to users. Nonetheless, early research in adversarial machine learning has unveiled that exposing model explanations can render victim models more susceptible to attacks. This is the first study to investigate and compare the impact of concept-based explanations on the privacy of Deep Learning based AI models in the context of biomedical image analysis. An extensive privacy benchmark is conducted on three different state-of-the-art model architectures (ResNet50, NFNet, ConvNeXt) trained on two biomedical (ISIC and EyePACS) and one synthetic dataset (SCDB). The success of membership inference attacks while exposing varying degrees of attribution-based and concept-based explanations is systematically compared. The findings indicate that, in theory, concept-based explanations can potentially increase the vulnerability of a private AI system by up to 16% compared to attributions in the baseline setting. However, it is demonstrated that, in more realistic attack scenarios, the threat posed by explanations is negligible in practice. Furthermore, actionable recommendations are provided to ensure the safe deployment of concept-based XAI systems. In addition, the impact of differential privacy (DP) on the quality of concept-based explanations is explored, revealing that while negatively influencing the explanation ability, DP can have an adverse effect on the models' privacy.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC10356902PMC
http://dx.doi.org/10.3389/fbinf.2023.1194993DOI Listing

Publication Analysis

Top Keywords

concept-based explanations
20
explanations
8
image analysis
8
concept-based
7
privacy
5
translating theory
4
theory practice
4
practice assessing
4
assessing privacy
4
privacy implications
4

Similar Publications

Article Synopsis
  • Deep learning has become the top choice for computer vision and decision-making tasks in the last decade, but its complex models raise concerns about fairness and understanding their decisions.
  • Existing methods provide insights into how low-level features relate to outcomes but struggle to connect these to higher-level human reasoning, leading to irrelevant concept extraction and limited local explanations.
  • The study introduces CA-SoftNet, a two-part model blending fast pattern recognition and clear reasoning, which shows improved accuracy on diverse datasets and offers a better way to extract and explain important concepts in a human-understandable manner.
View Article and Find Full Text PDF

MulCPred: Learning Multi-Modal Concepts for Explainable Pedestrian Action Prediction.

Sensors (Basel)

October 2024

Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-ku, Nagoya 464-8601, Japan.

Article Synopsis
  • The paper introduces MulCPred, a new framework designed to provide explainable predictions for pedestrian action, which is essential for applications like autonomous driving.
  • It addresses limitations of existing methods by using a linear aggregator for multi-modal concept integration, a channel-wise recalibration module for focusing on detailed input areas, and a regularization loss to capture diverse patterns.
  • Evaluation on various datasets shows that MulCPred enhances the explainability of predictions without significantly harming accuracy, and by filtering out unrecognizable concepts, it improves performance across different datasets.
View Article and Find Full Text PDF

Concept-based AI interpretability in physiological time-series data: Example of abnormality detection in electroencephalography.

Comput Methods Programs Biomed

December 2024

Institute for Biomedical Informatics, Faculty of Medicine, University Hospital Cologne, University of Cologne, Cologne, Germany.

Article Synopsis
  • Despite advancements in deep learning, its complex and opaque nature hinders widespread clinical adoption, prompting interest in concept-based interpretability, specifically using techniques like Testing with Concept Activation Vectors (TCAV).
  • This study applies TCAV to abnormality detection in electroencephalography (EEG), utilizing the XceptionTime model on multi-channel physiological data to enhance interpretability and analyze concepts linked to EEG pathologies.
  • The results indicate that TCAV scores align with clinical expectations, demonstrating its potential for improving interpretability in deep learning models and identifying biases in medical data.
View Article and Find Full Text PDF

Existing deep learning methods have achieved remarkable results in diagnosing retinal diseases, showcasing the potential of advanced AI in ophthalmology. However, the black-box nature of these methods obscures the decision-making process, compromising their trustworthiness and acceptability. Inspired by the concept-based approaches and recognizing the intrinsic correlation between retinal lesions and diseases, we regard retinal lesions as concepts and propose an inherently interpretable framework designed to enhance both the performance and explainability of diagnostic models.

View Article and Find Full Text PDF

End-to-end deep learning models are increasingly applied to safety-critical human activity recognition (HAR) applications, e.g., healthcare monitoring and smart home control, to reduce developer burden and increase the performance and robustness of prediction models.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!