Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
Constrained Counterfactual Explanations for Temporal Data
Stockholms universitet, Samhällsvetenskapliga fakulteten, Institutionen för data- och systemvetenskap.ORCID-id: 0000-0002-8575-421x
2024 (Engelska)Doktorsavhandling, sammanläggning (Övrigt vetenskapligt)
Abstract [en]

Recent advancements in machine learning models for temporal data have demonstrated high performance in predictive tasks like time series prediction and event sequence classification, yet these models often remain opaque. Counterfactual explanations offer actionable insights into these opaque models by suggesting input modifications to achieve desired predictive outcomes. In the context of explainable machine learning methods, there is a challenge in applying counterfactual explanation techniques to temporal data, as most previous research has focused on image or tabular data classification. Moreover, there is a growing need to extend counterfactual constraints to critical domains like healthcare, where it is crucial to incorporate clinical considerations.

To address these challenges, this thesis proposes novel machine learning models to generate counterfactual explanations for temporal data prediction, along with incorporating additional counterfactual constraints. In particular, this thesis focuses on three types of predictive models: (1) event sequence classification, (2) time series classification, and (3) time series forecasting. Furthermore, the integration of local temporal constraints and domain-specific constraints is proposed to emphasize the importance of temporal features and the relevance of application domains through extensive experimentation. 

This thesis is organized into three parts. The first part presents a counterfactual explanation method for medical event sequences, using style-transfer techniques and incorporating additional medical knowledge in modelling. The second part of the thesis focuses on univariate time series classification, proposing a novel solution that utilizes either latent representation or feature space perturbations, additionally incorporating temporal constraints to guide the counterfactual generation. The third part introduces the problem of counterfactual explanations for time series forecasting, proposes a gradient-based method, and extends to integrating domain-specific constraints for diabetes patients. The conclusion of this thesis summarizes the empirical findings and discusses future directions for applying counterfactual methods in real-world scenarios.

Ort, förlag, år, upplaga, sidor
Stockholm: Department of Computer and Systems Sciences, Stockholm University , 2024. , s. 84
Serie
Report Series / Department of Computer & Systems Sciences, ISSN 1101-8526 ; 24-015
Nyckelord [en]
Counterfactual explanations; Deep learning; Explainable machine learning; Healthcare
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
data- och systemvetenskap
Identifikatorer
URN: urn:nbn:se:su:diva-234540ISBN: 978-91-8014-979-2 (tryckt)ISBN: 978-91-8014-980-8 (digital)OAI: oai:DiVA.org:su-234540DiVA, id: diva2:1906268
Disputation
2024-12-04, L50, NOD-huset, Borgarfjordsgatan 12, Kista, Stockholm, 09:00 (Engelska)
Opponent
Handledare
Tillgänglig från: 2014-11-11 Skapad: 2024-10-16 Senast uppdaterad: 2024-10-29Bibliografiskt granskad
Delarbeten
1. Counterfactual Explanations for Survival Prediction of Cardiovascular ICU Patients
Öppna denna publikation i ny flik eller fönster >>Counterfactual Explanations for Survival Prediction of Cardiovascular ICU Patients
2021 (Engelska)Ingår i: Artificial Intelligence in Medicine: 19th International Conference on Artificial Intelligence in Medicine, AIME 2021, Virtual Event, June 15–18, 2021, Proceedings / [ed] Allan Tucker; Pedro Henriques Abreu; Jaime Cardoso; Pedro Pereira Rodrigues; David Riaño, Cham: Springer, 2021, s. 338-348Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

In recent years, machine learning methods have been rapidly implemented in the medical domain. However, current state-of-the-art methods usually produce opaque, black-box models. To address the lack of model transparency, substantial attention has been given to develop interpretable machine learning methods. In the medical domain, counterfactuals can provide example-based explanations for predictions, and show practitioners the modifications required to change a prediction from an undesired to a desired state. In this paper, we propose a counterfactual explanation solution for predicting the survival of cardiovascular ICU patients, by representing their electronic health record as a sequence of medical events, and generating counterfactuals by adopting and employing a text style-transfer technique. Experimental results on the MIMIC-III dataset strongly suggest that text style-transfer methods can be effectively adapted for the problem of counterfactual explanations in healthcare applications and can achieve competitive performance in terms of counterfactual validity, BLEU-4 and local outlier metrics. 

Ort, förlag, år, upplaga, sidor
Cham: Springer, 2021
Serie
Lecture Notes in Artificial Intelligence, ISSN 0302-9743, E-ISSN 1611-3349 ; 12721
Nyckelord
Counterfactual explanations, Survival prediction, Explainable models, Deep learning
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
data- och systemvetenskap
Identifikatorer
urn:nbn:se:su:diva-200373 (URN)10.1007/978-3-030-77211-6_38 (DOI)978-3-030-77210-9 (ISBN)978-3-030-77211-6 (ISBN)
Konferens
19th International Conference on Artificial Intelligence in Medicine (AIME 2021), virtual, June 15-18, 2021
Tillgänglig från: 2022-01-04 Skapad: 2022-01-04 Senast uppdaterad: 2024-10-16Bibliografiskt granskad
2. Style-transfer counterfactual explanations: An application to mortality prevention of ICU patients
Öppna denna publikation i ny flik eller fönster >>Style-transfer counterfactual explanations: An application to mortality prevention of ICU patients
2023 (Engelska)Ingår i: Artificial Intelligence in Medicine, ISSN 0933-3657, E-ISSN 1873-2860, Vol. 135, artikel-id 102457Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

In recent years, machine learning methods have been rapidly adopted in the medical domain. However, current state-of-the-art medical mining methods usually produce opaque, black-box models. To address the lack of model transparency, substantial attention has been given to developing interpretable machine learning models. In the medical domain, counterfactuals can provide example-based explanations for predictions, and show practitioners the modifications required to change a prediction from an undesired to a desired state. In this paper, we propose a counterfactual solution MedSeqCF for preventing the mortality of three cohorts of ICU patients, by representing their electronic health records as medical event sequences, and generating counterfactuals by adopting and employing a text style-transfer technique. We propose three model augmentations for MedSeqCF to integrate additional medical knowledge for generating more trustworthy counterfactuals. Experimental results on the MIMIC-III dataset strongly suggest that augmented style-transfer methods can be effectively adapted for the problem of counterfactual explanations in healthcare applications and can further improve the model performance in terms of validity, BLEU-4, local outlier factor, and edit distance. In addition, our qualitative analysis of the results by consultation with medical experts suggests that our style-transfer solutions can generate clinically relevant and actionable counterfactual explanations.

Nationell ämneskategori
Systemvetenskap, informationssystem och informatik
Forskningsämne
data- och systemvetenskap
Identifikatorer
urn:nbn:se:su:diva-212771 (URN)10.1016/j.artmed.2022.102457 (DOI)000897143800009 ()36628793 (PubMedID)2-s2.0-85143973748 (Scopus ID)
Tillgänglig från: 2022-12-12 Skapad: 2022-12-12 Senast uppdaterad: 2024-10-16Bibliografiskt granskad
3. Learning Time Series Counterfactuals via Latent Space Representations
Öppna denna publikation i ny flik eller fönster >>Learning Time Series Counterfactuals via Latent Space Representations
2021 (Engelska)Ingår i: Discovery Science: 24th International Conference, DS 2021, Halifax, NS, Canada, October 11–13, 2021, Proceedings / [ed] Carlos Soares; Luis Torgo, Springer , 2021, s. 369-384Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Counterfactual explanations can provide sample-based explanations of features required to modify from the original sample to change the classification result from an undesired state to a desired state; hence it provides interpretability of the model. Previous work of LatentCF presents an algorithm for image data that employs auto-encoder models to directly transform original samples into counterfactuals in a latent space representation. In our paper, we adapt the approach to time series classification and propose an improved algorithm named LatentCF++ which introduces additional constraints in the counterfactual generation process. We conduct an extensive experiment on a total of 40 datasets from the UCR archive, comparing to current state-of-the-art methods. Based on our evaluation metrics, we show that the LatentCF++ framework can with high probability generate valid counterfactuals and achieve comparable explanations to current state-of-the-art. Our proposed approach can also generate counterfactuals that are considerably closer to the decision boundary in terms of margin difference.

Ort, förlag, år, upplaga, sidor
Springer, 2021
Serie
Lecture Notes in Computer Science (LNCS), ISSN 0302-9743, E-ISSN 1611-3349
Nyckelord
Time series classification, Interpretability, Counterfactual explanations, Deep learning
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
data- och systemvetenskap
Identifikatorer
urn:nbn:se:su:diva-200497 (URN)10.1007/978-3-030-88942-5_29 (DOI)978-3-030-88941-8 (ISBN)978-3-030-88942-5 (ISBN)
Konferens
International Conference, DS 2021, Halifax, NS, Canada, October 11–13, 2021
Tillgänglig från: 2022-01-06 Skapad: 2022-01-06 Senast uppdaterad: 2024-10-16Bibliografiskt granskad
4. Glacier: guided locally constrained counterfactual explanations for time series classification
Öppna denna publikation i ny flik eller fönster >>Glacier: guided locally constrained counterfactual explanations for time series classification
Visa övriga...
2024 (Engelska)Ingår i: Machine Learning, ISSN 0885-6125, E-ISSN 1573-0565, Vol. 113, s. 4639-4669Artikel i tidskrift (Refereegranskat) Published
Abstract [en]

In machine learning applications, there is a need to obtain predictive models of high performance and, most importantly, to allow end-users and practitioners to understand and act on their predictions. One way to obtain such understanding is via counterfactuals, that provide sample-based explanations in the form of recommendations on which features need to be modified from a test example so that the classification outcome of a given classifier changes from an undesired outcome to a desired one. This paper focuses on the domain of time series classification, more specifically, on defining counterfactual explanations for univariate time series. We propose Glacier, a model-agnostic method for generating locally-constrained counterfactual explanations for time series classification using gradient search either on the original space or on a latent space that is learned through an auto-encoder. An additional flexibility of our method is the inclusion of constraints on the counterfactual generation process that favour applying changes to particular time series points or segments while discouraging changing others. The main purpose of these constraints is to ensure more reliable counterfactuals, while increasing the efficiency of the counterfactual generation process. Two particular types of constraints are considered, i.e., example-specific constraints and global constraints. We conduct extensive experiments on 40 datasets from the UCR archive, comparing different instantiations of Glacier against three competitors. Our findings suggest that Glacier outperforms the three competitors in terms of two common metrics for counterfactuals, i.e., proximity and compactness. Moreover, Glacier obtains comparable counterfactual validity compared to the best of the three competitors. Finally, when comparing the unconstrained variant of Glacier to the constraint-based variants, we conclude that the inclusion of example-specific and global constraints yields a good performance while demonstrating the trade-off between the different metrics.

Nyckelord
Time series classification, Interpretability, Counterfactual explanation, s Deep learning
Nationell ämneskategori
Annan data- och informationsvetenskap
Forskningsämne
data- och systemvetenskap
Identifikatorer
urn:nbn:se:su:diva-227717 (URN)10.1007/s10994-023-06502-x (DOI)001181943800001 ()2-s2.0-85187677577 (Scopus ID)
Tillgänglig från: 2024-03-26 Skapad: 2024-03-26 Senast uppdaterad: 2024-10-16Bibliografiskt granskad
5. Counterfactual Explanations for Time Series Forecasting
Öppna denna publikation i ny flik eller fönster >>Counterfactual Explanations for Time Series Forecasting
2024 (Engelska)Ingår i: 2023 IEEE International Conference on Data Mining (ICDM), IEEE conference proceedings , 2024, s. 1391-1396Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Among recent developments in time series forecasting methods, deep forecasting models have gained popularity as they can utilize hidden feature patterns in time series to improve forecasting performance. Nevertheless, the majority of current deep forecasting models are opaque, hence making it challenging to interpret the results. While counterfactual explanations have been extensively employed as a post-hoc approach for explaining classification models, their application to forecasting models still remains underexplored. In this paper, we formulate the novel problem of counterfactual generation for time series forecasting, and propose an algorithm, called ForecastCF, that solves the problem by applying gradient-based perturbations to the original time series. The perturbations are further guided by imposing constraints to the forecasted values. We experimentally evaluate ForecastCF using four state-of-the-art deep model architectures and compare to two baselines. ForecastCF outperforms the baselines in terms of counterfactual validity and data manifold closeness, while generating meaningful and relevant counterfactuals for various forecasting tasks.

Ort, förlag, år, upplaga, sidor
IEEE conference proceedings, 2024
Serie
IEEE International Conference on Data Mining. Proceedings, ISSN 1550-4786, E-ISSN 2374-8486
Nyckelord
Time series forecasting, Counterfactual explanations, Model interpretability, Deep learning
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
data- och systemvetenskap
Identifikatorer
urn:nbn:se:su:diva-226602 (URN)10.1109/ICDM58522.2023.00180 (DOI)001165180100171 ()2-s2.0-85185401353 (Scopus ID)979-8-3503-0788-7 (ISBN)
Konferens
IEEE International Conference on Data Mining (ICDM), 1-4 December 2023, Shanghai, China.
Tillgänglig från: 2024-02-14 Skapad: 2024-02-14 Senast uppdaterad: 2024-11-14Bibliografiskt granskad
6. COMET: Constrained Counterfactual Explanations for Patient Glucose Multivariate Forecasting
Öppna denna publikation i ny flik eller fönster >>COMET: Constrained Counterfactual Explanations for Patient Glucose Multivariate Forecasting
2024 (Engelska)Ingår i: Annual IEEE Symposium on Computer-Based Medical Systems: 2024 IEEE 37th International Symposium on Computer-Based Medical Systems (CBMS), 26-28 June 2024, IEEE (Institute of Electrical and Electronics Engineers) , 2024, s. 502-507Konferensbidrag, Publicerat paper (Refereegranskat)
Abstract [en]

Applying deep learning models for healthcare-related forecasting applications has been widely adopted, such as leveraging glucose monitoring data of diabetes patients to predict hyperglycaemic or hypoglycaemic events. However, most deep learning models are considered black-boxes; hence, the model predictions are not interpretable and may not offer actionable insights into medical practitioners’ decisions. Previous work has shown that counterfactual explanations can be applied in forecasting tasks by suggesting counterfactual changes in time series inputs to achieve the desired forecasting outcome. This study proposes a generalized multivariate forecasting setup of counterfactual generation by introducing a novel approach, COMET, which imposes three domain-specific constraint mechanisms to provide counterfactual explanations for glucose forecasting. Moreover, we conduct the experimental evaluation using two diabetes patient datasets to demonstrate the effectiveness of our proposed approach in generating realistic counterfactual changes in comparison with a baseline approach. Our qualitative analysis evaluates examples to validate that the counterfactual samples are clinically relevant and can effectively lead the patients to achieve a normal range of predicted glucose levels by suggesting changes to the treatment variables.

Ort, förlag, år, upplaga, sidor
IEEE (Institute of Electrical and Electronics Engineers), 2024
Serie
IEEE International Symposium on Computer-Based Medical Systems, ISSN 2372-918X, E-ISSN 2372-9198
Nyckelord
Comet, Deep learning, Patents, Time series analysis, Predictive models, Glucose, Diabetes, time series forecasting, blood glucose prediction, counterfactual explanations, deep learning
Nationell ämneskategori
Datavetenskap (datalogi)
Forskningsämne
data- och systemvetenskap
Identifikatorer
urn:nbn:se:su:diva-233744 (URN)10.1109/CBMS61543.2024.00089 (DOI)001284700700038 ()2-s2.0-85200437241 (Scopus ID)
Konferens
2024 IEEE 37th International Symposium on Computer-Based Medical Systems (CBMS), 26-28 June 2024, Guadalajara, Mexico.
Tillgänglig från: 2024-09-24 Skapad: 2024-09-24 Senast uppdaterad: 2024-10-16Bibliografiskt granskad

Open Access i DiVA

Constrained Counterfactual Explanations for Temporal Data(4889 kB)404 nedladdningar
Filinformation
Filnamn FULLTEXT03.pdfFilstorlek 4889 kBChecksumma SHA-512
07ff88cd5439eff75309f0e03819c26dd7e7b964871da4e179438e82b7cfedc3c3394fea5cf7c6b3854ddb8469f857d743439188fcf8af27c5f76cb95e055db7
Typ fulltextMimetyp application/pdf

Person

Wang, Zhendong

Sök vidare i DiVA

Av författaren/redaktören
Wang, Zhendong
Av organisationen
Institutionen för data- och systemvetenskap
Datavetenskap (datalogi)

Sök vidare utanför DiVA

GoogleGoogle Scholar
Totalt: 409 nedladdningar
Antalet nedladdningar är summan av nedladdningar för alla fulltexter. Det kan inkludera t.ex tidigare versioner som nu inte längre är tillgängliga.

isbn
urn-nbn

Altmetricpoäng

isbn
urn-nbn
Totalt: 1034 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf