This repository provides source codes to reproduce experiments described in Model-based Data-Complexity Estimator for Deep Learning Systems.
- Install poetry
- Install dependencies
$ poetry install
- Run the following command
$ poetry shell $ python ./onestop.py [path_to_your_model] [path_to_training_dataset] [layer_name] (--test [path_to_test_dataset]) (--tag [tag_to_identify_results])
- The calculated complexities are saved in ``results'' folder
- Run the following commannd
$ cd experiments $ poetry install $ poetry shell $ jupyter notebook
- Open jupyter notebook and run cells in order.
. ├─datasets │ ├─test │ └─training ├─docs ├─experiments │ ├─activation_traces │ │ ├─test │ │ └─training │ ├─nmf_info │ │ ├─base_weight │ │ └─model │ ├─pred_res │ └─results ├─intermediate_results │ ├─activation_traces │ └─nmf_information ├─results ├─src │ └─lib └─trained_models
Datasets used in the experiments (.npz files)
Files for GitHub Pages showing supplemental materials
Resource to conduct experiments in the paper.
Activation Traces used in the experiments.
Since the file size is too large, we just show hash values of the files in this repo.
If you need the original files, please contact us.
- base_weight
Feature Matrix obtained in the experiments (.npz files) - model
NMF that fit to the activation traces of inputs in each training dataset (.pkl files)
AT and NMF information is stored if you calculate complexities on your model and datasets.
Calculated complexities will be placed.
Source codes to calculate complexities.
Trained model used in the experiments (.h5 files)