Skip to content

Commit

Permalink
STY: Re-run make specs; change in black
Browse files Browse the repository at this point in the history
  • Loading branch information
effigies committed Sep 17, 2020
1 parent 97ce446 commit 8c4a8b0
Show file tree
Hide file tree
Showing 833 changed files with 40,312 additions and 10,893 deletions.
72 changes: 55 additions & 17 deletions nipype/algorithms/tests/test_auto_ACompCor.py
Expand Up @@ -4,24 +4,56 @@

def test_ACompCor_inputs():
input_map = dict(
components_file=dict(usedefault=True,),
failure_mode=dict(usedefault=True,),
components_file=dict(
usedefault=True,
),
failure_mode=dict(
usedefault=True,
),
header_prefix=dict(),
high_pass_cutoff=dict(usedefault=True,),
ignore_initial_volumes=dict(usedefault=True,),
high_pass_cutoff=dict(
usedefault=True,
),
ignore_initial_volumes=dict(
usedefault=True,
),
mask_files=dict(),
mask_index=dict(requires=["mask_files"], xor=["merge_method"],),
mask_index=dict(
requires=["mask_files"],
xor=["merge_method"],
),
mask_names=dict(),
merge_method=dict(requires=["mask_files"], xor=["mask_index"],),
num_components=dict(xor=["variance_threshold"],),
pre_filter=dict(usedefault=True,),
realigned_file=dict(extensions=None, mandatory=True,),
regress_poly_degree=dict(usedefault=True,),
merge_method=dict(
requires=["mask_files"],
xor=["mask_index"],
),
num_components=dict(
xor=["variance_threshold"],
),
pre_filter=dict(
usedefault=True,
),
realigned_file=dict(
extensions=None,
mandatory=True,
),
regress_poly_degree=dict(
usedefault=True,
),
repetition_time=dict(),
save_metadata=dict(usedefault=True,),
save_pre_filter=dict(usedefault=True,),
use_regress_poly=dict(deprecated="0.15.0", new_name="pre_filter",),
variance_threshold=dict(xor=["num_components"],),
save_metadata=dict(
usedefault=True,
),
save_pre_filter=dict(
usedefault=True,
),
use_regress_poly=dict(
deprecated="0.15.0",
new_name="pre_filter",
),
variance_threshold=dict(
xor=["num_components"],
),
)
inputs = ACompCor.input_spec()

Expand All @@ -32,9 +64,15 @@ def test_ACompCor_inputs():

def test_ACompCor_outputs():
output_map = dict(
components_file=dict(extensions=None,),
metadata_file=dict(extensions=None,),
pre_filter_file=dict(extensions=None,),
components_file=dict(
extensions=None,
),
metadata_file=dict(
extensions=None,
),
pre_filter_file=dict(
extensions=None,
),
)
outputs = ACompCor.output_spec()

Expand Down
21 changes: 17 additions & 4 deletions nipype/algorithms/tests/test_auto_ActivationCount.py
Expand Up @@ -3,7 +3,14 @@


def test_ActivationCount_inputs():
input_map = dict(in_files=dict(mandatory=True,), threshold=dict(mandatory=True,),)
input_map = dict(
in_files=dict(
mandatory=True,
),
threshold=dict(
mandatory=True,
),
)
inputs = ActivationCount.input_spec()

for key, metadata in list(input_map.items()):
Expand All @@ -13,9 +20,15 @@ def test_ActivationCount_inputs():

def test_ActivationCount_outputs():
output_map = dict(
acm_neg=dict(extensions=None,),
acm_pos=dict(extensions=None,),
out_file=dict(extensions=None,),
acm_neg=dict(
extensions=None,
),
acm_pos=dict(
extensions=None,
),
out_file=dict(
extensions=None,
),
)
outputs = ActivationCount.output_spec()

Expand Down
16 changes: 13 additions & 3 deletions nipype/algorithms/tests/test_auto_AddCSVColumn.py
Expand Up @@ -6,8 +6,14 @@ def test_AddCSVColumn_inputs():
input_map = dict(
extra_column_heading=dict(),
extra_field=dict(),
in_file=dict(extensions=None, mandatory=True,),
out_file=dict(extensions=None, usedefault=True,),
in_file=dict(
extensions=None,
mandatory=True,
),
out_file=dict(
extensions=None,
usedefault=True,
),
)
inputs = AddCSVColumn.input_spec()

Expand All @@ -17,7 +23,11 @@ def test_AddCSVColumn_inputs():


def test_AddCSVColumn_outputs():
output_map = dict(csv_file=dict(extensions=None,),)
output_map = dict(
csv_file=dict(
extensions=None,
),
)
outputs = AddCSVColumn.output_spec()

for key, metadata in list(output_map.items()):
Expand Down
14 changes: 12 additions & 2 deletions nipype/algorithms/tests/test_auto_AddCSVRow.py
Expand Up @@ -4,7 +4,13 @@

def test_AddCSVRow_inputs():
input_map = dict(
_outputs=dict(usedefault=True,), in_file=dict(extensions=None, mandatory=True,),
_outputs=dict(
usedefault=True,
),
in_file=dict(
extensions=None,
mandatory=True,
),
)
inputs = AddCSVRow.input_spec()

Expand All @@ -14,7 +20,11 @@ def test_AddCSVRow_inputs():


def test_AddCSVRow_outputs():
output_map = dict(csv_file=dict(extensions=None,),)
output_map = dict(
csv_file=dict(
extensions=None,
),
)
outputs = AddCSVRow.output_spec()

for key, metadata in list(output_map.items()):
Expand Down
33 changes: 26 additions & 7 deletions nipype/algorithms/tests/test_auto_AddNoise.py
Expand Up @@ -4,12 +4,27 @@

def test_AddNoise_inputs():
input_map = dict(
bg_dist=dict(mandatory=True, usedefault=True,),
dist=dict(mandatory=True, usedefault=True,),
in_file=dict(extensions=None, mandatory=True,),
in_mask=dict(extensions=None,),
out_file=dict(extensions=None,),
snr=dict(usedefault=True,),
bg_dist=dict(
mandatory=True,
usedefault=True,
),
dist=dict(
mandatory=True,
usedefault=True,
),
in_file=dict(
extensions=None,
mandatory=True,
),
in_mask=dict(
extensions=None,
),
out_file=dict(
extensions=None,
),
snr=dict(
usedefault=True,
),
)
inputs = AddNoise.input_spec()

Expand All @@ -19,7 +34,11 @@ def test_AddNoise_inputs():


def test_AddNoise_outputs():
output_map = dict(out_file=dict(extensions=None,),)
output_map = dict(
out_file=dict(
extensions=None,
),
)
outputs = AddNoise.output_spec()

for key, metadata in list(output_map.items()):
Expand Down
70 changes: 53 additions & 17 deletions nipype/algorithms/tests/test_auto_ArtifactDetect.py
Expand Up @@ -4,25 +4,61 @@

def test_ArtifactDetect_inputs():
input_map = dict(
bound_by_brainmask=dict(usedefault=True,),
global_threshold=dict(usedefault=True,),
intersect_mask=dict(usedefault=True,),
mask_file=dict(extensions=None,),
bound_by_brainmask=dict(
usedefault=True,
),
global_threshold=dict(
usedefault=True,
),
intersect_mask=dict(
usedefault=True,
),
mask_file=dict(
extensions=None,
),
mask_threshold=dict(),
mask_type=dict(mandatory=True,),
mask_type=dict(
mandatory=True,
),
norm_threshold=dict(
mandatory=True, xor=["rotation_threshold", "translation_threshold"],
),
parameter_source=dict(mandatory=True,),
plot_type=dict(usedefault=True,),
realigned_files=dict(mandatory=True,),
realignment_parameters=dict(mandatory=True,),
rotation_threshold=dict(mandatory=True, xor=["norm_threshold"],),
save_plot=dict(usedefault=True,),
translation_threshold=dict(mandatory=True, xor=["norm_threshold"],),
use_differences=dict(maxlen=2, minlen=2, usedefault=True,),
use_norm=dict(requires=["norm_threshold"], usedefault=True,),
zintensity_threshold=dict(mandatory=True,),
mandatory=True,
xor=["rotation_threshold", "translation_threshold"],
),
parameter_source=dict(
mandatory=True,
),
plot_type=dict(
usedefault=True,
),
realigned_files=dict(
mandatory=True,
),
realignment_parameters=dict(
mandatory=True,
),
rotation_threshold=dict(
mandatory=True,
xor=["norm_threshold"],
),
save_plot=dict(
usedefault=True,
),
translation_threshold=dict(
mandatory=True,
xor=["norm_threshold"],
),
use_differences=dict(
maxlen=2,
minlen=2,
usedefault=True,
),
use_norm=dict(
requires=["norm_threshold"],
usedefault=True,
),
zintensity_threshold=dict(
mandatory=True,
),
)
inputs = ArtifactDetect.input_spec()

Expand Down
10 changes: 8 additions & 2 deletions nipype/algorithms/tests/test_auto_CalculateMedian.py
Expand Up @@ -4,7 +4,11 @@

def test_CalculateMedian_inputs():
input_map = dict(
in_files=dict(), median_file=dict(), median_per_file=dict(usedefault=True,),
in_files=dict(),
median_file=dict(),
median_per_file=dict(
usedefault=True,
),
)
inputs = CalculateMedian.input_spec()

Expand All @@ -14,7 +18,9 @@ def test_CalculateMedian_inputs():


def test_CalculateMedian_outputs():
output_map = dict(median_files=dict(),)
output_map = dict(
median_files=dict(),
)
outputs = CalculateMedian.output_spec()

for key, metadata in list(output_map.items()):
Expand Down
13 changes: 10 additions & 3 deletions nipype/algorithms/tests/test_auto_CalculateNormalizedMoments.py
Expand Up @@ -4,8 +4,13 @@

def test_CalculateNormalizedMoments_inputs():
input_map = dict(
moment=dict(mandatory=True,),
timeseries_file=dict(extensions=None, mandatory=True,),
moment=dict(
mandatory=True,
),
timeseries_file=dict(
extensions=None,
mandatory=True,
),
)
inputs = CalculateNormalizedMoments.input_spec()

Expand All @@ -15,7 +20,9 @@ def test_CalculateNormalizedMoments_inputs():


def test_CalculateNormalizedMoments_outputs():
output_map = dict(moments=dict(),)
output_map = dict(
moments=dict(),
)
outputs = CalculateNormalizedMoments.output_spec()

for key, metadata in list(output_map.items()):
Expand Down

0 comments on commit 8c4a8b0

Please sign in to comment.