Skip to content

Commit

Permalink
Remove obsolete generate_experiment_id.py (Closes #111)
Browse files Browse the repository at this point in the history
  • Loading branch information
dweindl committed Dec 4, 2019
1 parent 5c399c9 commit 7eec68e
Show file tree
Hide file tree
Showing 4 changed files with 1 addition and 141 deletions.
2 changes: 1 addition & 1 deletion CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ Library:
* Setup documentation for RTD (#161)
* Handle None in petab.core.split_parameter_replacement_list (Closes #121)
* Fix(lint) correct handling of optional columns. Check before access.

* Remove obsolete generate_experiment_id.py (Closes #111) #166

### 0.0.0a16 and earlier

Expand Down
1 change: 0 additions & 1 deletion petab/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,5 +4,4 @@
from .lint import * # noqa: F403, F401
from .sbml import * # noqa: F403, F401
from .parameter_mapping import * # noqa: F403, F401
from .generate_experiment_id import generate_experiment_id # noqa: F403, F401
from .version import __version__ # noqa: F401
90 changes: 0 additions & 90 deletions petab/generate_experiment_id.py

This file was deleted.

49 changes: 0 additions & 49 deletions tests/test_visualization.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,5 @@
import pandas as pd
import pytest
import warnings
from petab import generate_experiment_id # noqa: E402
from petab.visualize import (plot_data_and_simulation,
plot_measurements_by_observable)

Expand Down Expand Up @@ -227,50 +225,3 @@ def test_simple_visualization(data_file_Fujita,
plot_measurements_by_observable(data_file_Fujita, condition_file_Fujita)
plot_measurements_by_observable(data_file_Fujita, condition_file_Fujita,
plotted_noise='provided')


def test_generate_experimentId_no_empty():
data_actual = {'observableParameters': ['obs1', 'obs1', 'obs2', 'obs3'],
'noiseParameters': ['noise1', 'noise1', 'noise2',
'noise3'],
'observableTransformation': ['log10', 'log10', 'log10',
'log10']}
measurement_data = pd.DataFrame(data_actual)
measurement_data_actual = generate_experiment_id(measurement_data)
data_expect = {'observableParameters': ['obs1', 'obs1', 'obs2', 'obs3'],
'noiseParameters': ['noise1', 'noise1', 'noise2',
'noise3'],
'observableTransformation': ['log10', 'log10', 'log10',
'log10'],
'experimentId': ['experiment_1', 'experiment_1',
'experiment_2', 'experiment_3']}
measurement_data_expect = pd.DataFrame(data_expect)
assert all(measurement_data_actual == measurement_data_expect)


def test_generate_experiment_id_empty():
data_actual = {'observableParameters': ['obs1', 'obs1', 'obs2', 'obs3',
float('nan'), 1, 1, 2, 2],
'noiseParameters': ['noise1', 'noise1', 'noise2',
'noise3', 'noise3', 'noise4',
'noise4', 'noise4', 'noise5'],
'observableTransformation': ['log10', 'log10', 'log10',
'log10', 'log10', 'log10',
'log10', 'log10', 'log10']}
measurement_data = pd.DataFrame(data_actual)
measurement_data_actual = generate_experiment_id(measurement_data)
data_expect = {'observableParameters': ['obs1', 'obs1', 'obs2', 'obs3',
float('nan'), 1, 1, 2, 2],
'noiseParameters': ['noise1', 'noise1', 'noise2',
'noise3', 'noise3', 'noise4',
'noise4', 'noise4', 'noise5'],
'observableTransformation': ['log10', 'log10', 'log10',
'log10', 'log10', 'log10',
'log10', 'log10', 'log10'],
'experimentId': ['experiment_1', 'experiment_1',
'experiment_2', 'experiment_3',
'experiment_4', 'experiment_5',
'experiment_5', 'experiment_6',
'experiment_7']}
measurement_data_expect = pd.DataFrame(data_expect)
assert all(measurement_data_actual == measurement_data_expect)

0 comments on commit 7eec68e

Please sign in to comment.