Skip to content

How to automate feature generation process for large scale time series data using two python libraries tsfresh and Dask

License

Notifications You must be signed in to change notification settings

arnabbiswas1/feature_engineering_with_tsfresh_and_dask

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Automatic Feature Engineering for Large Scale Time Series Data Using tsfresh and Dask

The internet of things, digitized health care systems, financial markets, smart cities (etc.) are continuously generating time series data of different types, sizes and complexities. Time series data is different from non-temporal data. In time series data, observation at any instance of time depends on the observations from the past based on the underlying process. Often it contains noise and redundant information. To make things more complex, most of the traditional Machine Learning algorithms are developed for non-temporal data. Thus, extracting meaningful features from raw time series plays a major role. While there are features generic across different flavors/types of time series, there are features specific to different domains. As a result, feature engineering often demands familiarity with domain specific and/or signal processing algorithms making the process complicated.

This repository introduces to a Python library called tsfresh. tsfresh accelerates the feature engineering process by automatically generating 750+ of features for time series data.

However, if the size of the time series data is large, we start encountering two kinds problems:

  • Large execution time
  • Need for larger memory

This is where another Python framework Dask comes into picture. Dask parallelizes the feature extraction process of tsfresh. Also, by using out of core computing, it addresses the problem of larger than RAM dataset.

This content would be presented at a Python Conference, BelPy 2021. This content was used for a talk at PyData Montreal Jan 2021 meetup:

To gather basic understanding of dask, please refer to this repository.

How to setup the Python Environment for this project?

conda env create -f environment.yml
conda activate dask_tsfresh
jupyter labextension install dask-labextension
jupyter serverextension enable dask_labextension

How to get the data?

Data used in this project is from the Kaggle Competition "INGV - Volcanic Eruption Prediction"(https://www.kaggle.com/c/predict-volcanic-eruptions-ingv-oe). Please download the data from Kaggle and place into the data directory. Once done, the content of the data directory should have the following files & directories:

sample_submission.csv  
test/                  
train/                 
train.csv

About

How to automate feature generation process for large scale time series data using two python libraries tsfresh and Dask

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published