EXPLAINING TIME SERIES VIA CONTRASTIVE AND LOCALLY SPARSE PERTURBATIONS

  • Zichuan Liu
  • , Yingying Zhang
  • , Tianchun Wang
  • , Zefan Wang
  • , Dongsheng Luo
  • , Mengnan Du
  • , Min Wu
  • , Yi Wang
  • , Chunlin Chen
  • , Lunting Fan
  • , Qingsong Wen

Research output: Contribution to conferencePaperpeer-review

8 Scopus citations

Abstract

Explaining multivariate time series is a compound challenge, as it requires identifying important locations in the time series and matching complex temporal patterns. Although previous saliency-based methods addressed the challenges, their perturbation may not alleviate the distribution shift issue, which is inevitable especially in heterogeneous samples. We present ContraLSP, a locally sparse model that introduces counterfactual samples to build uninformative perturbations but keeps distribution using contrastive learning. Furthermore, we incorporate sample-specific sparse gates to generate more binary-skewed and smooth masks, which easily integrate temporal trends and select the salient features parsimoniously. Empirical studies on both synthetic and real-world datasets show that ContraLSP outperforms state-of-the-art models, demonstrating a substantial improvement in explanation quality for time series data. The source code is available at https://github.com/zichuan-liu/ContraLSP.

Original languageEnglish (US)
StatePublished - 2024
Event12th International Conference on Learning Representations, ICLR 2024 - Hybrid, Vienna, Austria
Duration: May 7 2024May 11 2024

Conference

Conference12th International Conference on Learning Representations, ICLR 2024
Country/TerritoryAustria
CityHybrid, Vienna
Period5/7/245/11/24

All Science Journal Classification (ASJC) codes

  • Language and Linguistics
  • Computer Science Applications
  • Education
  • Linguistics and Language

Fingerprint

Dive into the research topics of 'EXPLAINING TIME SERIES VIA CONTRASTIVE AND LOCALLY SPARSE PERTURBATIONS'. Together they form a unique fingerprint.

Cite this