Abstract

Interpretability has become a crucial component in the Machine Learning (ML) domain. This is particularly important in the context of medical and health applications, where the underlying reasons behind how an ML model makes a certain decision are as important as the decision itself for the experts. However, interpreting an ML model based on limited local data may potentially lead to inaccurate conclusions. On the other hand, centralized decision making and interpretability, by transferring the data to a centralized server, may raise privacy concerns due to the sensitivity of personal/medical data in such applications.

In this paper, we propose a federated interpretability scheme based on SHAP (SHapley Additive exPlanations) value and DeepLIFT (Deep Learning Important FeaTures) to interpret ML models, without sharing sensitive data and in a privacy-preserving fashion. Our proposed federated interpretability scheme is a decentralized framework for interpreting ML models, where data remains on local devices, and only values that do not directly describe the raw data are aggregated in a privacy-preserving fashion to interpret the model.
Original languageEnglish
Title of host publicationBig Data
PublisherIEEE - Institute of Electrical and Electronics Engineers Inc.
Publication statusAccepted/In press - 2024
EventIEEE International Conference on Big Data, BigData 2024 - Washington, United States
Duration: 2024 Dec 152024 Dec 18

Conference

ConferenceIEEE International Conference on Big Data, BigData 2024
Country/TerritoryUnited States
CityWashington
Period2024/12/152024/12/18

Subject classification (UKÄ)

  • Computer Systems

Free keywords

  • explainable machine learning
  • privacy-preserving
  • federated learning
  • epilepsy
  • seizure prediction
  • seizure Detection
  • EEG
  • ECG

Cite this