Skip to content

Code for the Paper Constructing Global Coherence Representations:Identifying Interpretability and Coherences ofTransformer Attention in Time Series Data. It is about creating coherence matrices which represent the attention from each symbol to each other symbol.

License

Notifications You must be signed in to change notification settings

cslab-hub/GlobalTimeSeriesCoherenceMatrices

Repository files navigation

GlobalTimeSeriesCoherenceMatrices

Description

Code for the Paper Constructing Global Coherence Representations:Identifying Interpretability and Coherences of Transformer Attention in Time Series Data. It is about creating coherence matrices which represent the attention from each symbol to each other symbol to enhance the understanding of the global class representation. It can be used as a classification model while providing a nice visualization and boost to the interpretability of the global classes

The project contains two Jupiter notebooks which provide the model from the publication and also includes the weights for the published results (in the "saves"-folder for 500 and 15 epochs). The saved weights need to be in the saved folder without a nested structure to be loaded.

  • GlobalTransformerInterpretation-ForReproducibility.ipynb: Contains the model from the paper for reproducibility.
  • GlobalTransformerInterpretation-WithFix.ipynb: Contains a fix; results are not compatible with the publication.

The code was tested on four datasets (linked below and included in the repository) and trains with a 5 fold cross-validation. Each fold trains 2 models and 10 coherence representations:

Models:

  • Normal input data
  • Symbolic data (SAX)

Coherence Representations:

  • Full Coherence Attention Matrices based on sum
  • Full Coherence Attention Matrices based on relative average
  • Column Reduced Coherence Attention Matrices based on sum
  • Column Reduced Coherence Attention Matrices based on relative average
  • Global Trend Matrix based on max of sum
  • Global Trend Matrix based on max of relative average
  • Global Trend Matrix based on average of sum
  • Global Trend Matrix based on average of relative average
  • Global Trend Matrix based on median of sum
  • Global Trend Matrix based on median of relative average

At the end of the notebook the coherence matrices can be analyzed with the given visualizations.

Dependencies

A list of all needed dependencies (other versions can work but are not guaranteed to do so):

python=3.7.3
tensorflow==2.2.0
tensorflow_addons==0.11.2
tensorflow_probability==0.7.0
seaborn==0.10.1
scipy==1.4.1
scikit-learn==0.23.2
pyts==0.11.0
pandas==1.0.0
numpy==1.18.5
matplotlib==3.3.1

Cite and publications

This code represents the used model for the following publication:
https://ieeexplore.ieee.org/abstract/document/9564126

If you use, build upon this work or if it helped in any other way, please cite the linked publication.

Datasets

Included datasets are:

http://www.timeseriesclassification.com/description.php?Dataset=SyntheticControl
http://www.timeseriesclassification.com/description.php?Dataset=ECG5000
http://www.timeseriesclassification.com/description.php?Dataset=Plane
http://www.timeseriesclassification.com/description.php?Dataset=PowerCons

About

Code for the Paper Constructing Global Coherence Representations:Identifying Interpretability and Coherences ofTransformer Attention in Time Series Data. It is about creating coherence matrices which represent the attention from each symbol to each other symbol.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages