Skip to content

Commit

Permalink
pep8
Browse files Browse the repository at this point in the history
  • Loading branch information
Alexis Wolfsdorf committed Oct 21, 2019
1 parent a765f11 commit 5881dfc
Show file tree
Hide file tree
Showing 10 changed files with 53 additions and 21 deletions.
3 changes: 2 additions & 1 deletion pydatajson/catalog_readme.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,8 @@
TEMPLATES_PATH = os.path.join(ABSOLUTE_PROJECT_DIR, "templates")


def generate_catalog_readme(_datajson, catalog, export_path=None, verify_ssl=True):
def generate_catalog_readme(_datajson, catalog,
export_path=None, verify_ssl=True):
"""Este método está para mantener retrocompatibilidad con versiones
anteriores. Se ignora el argumento _data_json."""
return generate_readme(catalog, export_path, verify_ssl=verify_ssl)
Expand Down
3 changes: 2 additions & 1 deletion pydatajson/core.py
Original file line number Diff line number Diff line change
Expand Up @@ -967,7 +967,8 @@ def generate_catalogs_indicators(self, catalogs=None,
catalogs = catalogs or self
return indicators.generate_catalogs_indicators(
catalogs, central_catalog, identifier_search=identifier_search,
validator=self.validator, broken_links=broken_links, verify_ssl=self.verify_ssl)
validator=self.validator, broken_links=broken_links,
verify_ssl=self.verify_ssl)

def _count_fields_recursive(self, dataset, fields):
"""Cuenta la información de campos optativos/recomendados/requeridos
Expand Down
9 changes: 7 additions & 2 deletions pydatajson/indicators.py
Original file line number Diff line number Diff line change
Expand Up @@ -300,7 +300,9 @@ def _generate_status_indicators(catalog, validator=None, verify_ssl=True):
'datasets_con_datos_pct': None
}
try:
generator = StatusIndicatorsGenerator(catalog, validator=validator, verify_ssl=verify_ssl)
generator = StatusIndicatorsGenerator(catalog,
validator=validator,
verify_ssl=verify_ssl)
except Exception as e:
msg = u'Error generando resumen del catálogo {}: {}'.format(
catalog['title'], str(e))
Expand Down Expand Up @@ -572,7 +574,10 @@ def _generate_valid_urls_indicators(catalog, validator=None, verify_ssl=True):

result = {}
try:
generator = StatusIndicatorsGenerator(catalog, validator=validator, verify_ssl=verify_ssl)
generator = \
StatusIndicatorsGenerator(catalog,
validator=validator,
verify_ssl=verify_ssl)
except Exception as e:
msg = u'Error generando resumen del catálogo {}: {}'.format(
catalog['title'], str(e))
Expand Down
4 changes: 3 additions & 1 deletion pydatajson/status_indicators_generator.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,8 @@
from pydatajson.readers import read_catalog
from pydatajson.reporting import generate_datasets_summary
from pydatajson.validators.distribution_download_urls_validator import DistributionDownloadUrlsValidator
from pydatajson.validators\
.distribution_download_urls_validator \
import DistributionDownloadUrlsValidator


class StatusIndicatorsGenerator(object):
Expand Down
14 changes: 10 additions & 4 deletions pydatajson/validation.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,11 +54,15 @@ def init_jsonschema_validator(self, schema_dir, schema_filename):
schema=schema, resolver=resolver, format_checker=format_checker)

def is_valid(self, catalog, broken_links=False, verify_ssl=True):
return not self._get_errors(catalog, broken_links=broken_links, verify_ssl=verify_ssl)
return not self._get_errors(catalog,
broken_links=broken_links,
verify_ssl=verify_ssl)

def validate_catalog(self, catalog, only_errors=False, broken_links=False, verify_ssl=True):
def validate_catalog(self, catalog, only_errors=False,
broken_links=False, verify_ssl=True):
default_response = self._default_response(catalog)
errors = self._get_errors(catalog, broken_links=broken_links, verify_ssl=verify_ssl)
errors = self._get_errors(catalog, broken_links=broken_links,
verify_ssl=verify_ssl)

response = default_response.copy()
for error in errors:
Expand Down Expand Up @@ -254,4 +258,6 @@ def validate_catalog(catalog, only_errors=False, fmt="dict",
else:
validator = Validator()

return validator.validate_catalog(catalog, only_errors, verify_ssl=verify_ssl)
return validator.validate_catalog(catalog,
only_errors,
verify_ssl=verify_ssl)
3 changes: 2 additions & 1 deletion pydatajson/validators/theme_ids_not_repeated_validator.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,8 @@ class ThemeIdsNotRepeatedValidator(SimpleValidator):

def validate(self):
if "themeTaxonomy" in self.catalog:
theme_ids = [theme["id"] for theme in self.catalog["themeTaxonomy"]]
theme_ids = [theme["id"]
for theme in self.catalog["themeTaxonomy"]]
dups = self._find_dups(theme_ids)
if len(dups) > 0:
yield ce.ThemeIdRepeated(dups)
Expand Down
3 changes: 2 additions & 1 deletion pydatajson/validators/url_validator.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,8 @@
import requests
from requests import RequestException, Timeout

from pydatajson.constants import EXCEPTION_STATUS_CODES, INVALID_STATUS_CODES_REGEX
from pydatajson.constants import EXCEPTION_STATUS_CODES, \
INVALID_STATUS_CODES_REGEX
from pydatajson.validators.simple_validator import SimpleValidator


Expand Down
3 changes: 2 additions & 1 deletion tests/test_core.py
Original file line number Diff line number Diff line change
Expand Up @@ -566,7 +566,8 @@ def test_generate_datasets_summary(self):
"several_datasets_for_harvest.json")

with mock.patch(
'pydatajson.validators.url_validator.UrlValidator.is_working_url',
'pydatajson.validators.url_validator'
'.UrlValidator.is_working_url',
return_value=(True, 200)):
actual = self.dj.generate_datasets_summary(catalog)

Expand Down
3 changes: 2 additions & 1 deletion tests/test_status_indicators_generator.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,8 @@ def get_sample(cls, sample_filename):
@classmethod
def setUpClass(cls):
with mock.patch(
'pydatajson.validators.url_validator.UrlValidator.is_working_url',
'pydatajson.validators'
'.url_validator.UrlValidator.is_working_url',
return_value=(True, 200)):
cls.gen_justicia = StatusIndicatorsGenerator(
cls.get_sample('catalogo_justicia.json'))
Expand Down
29 changes: 21 additions & 8 deletions tests/test_validators.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,10 +9,15 @@
import requests_mock
from requests import Timeout, ConnectionError

from pydatajson.validators.consistent_distribution_fields_validator import ConsistentDistributionFieldsValidator
from pydatajson.validators.distribution_urls_validator import DistributionUrlsValidator
from pydatajson.validators.landing_pages_validator import LandingPagesValidator
from pydatajson.validators.theme_ids_not_repeated_validator import ThemeIdsNotRepeatedValidator
from pydatajson.validators \
.consistent_distribution_fields_validator \
import ConsistentDistributionFieldsValidator
from pydatajson.validators.distribution_urls_validator \
import DistributionUrlsValidator
from pydatajson.validators.landing_pages_validator \
import LandingPagesValidator
from pydatajson.validators.theme_ids_not_repeated_validator \
import ThemeIdsNotRepeatedValidator
from pydatajson.validators.url_validator import UrlValidator

try:
Expand Down Expand Up @@ -96,7 +101,9 @@ def test_valid_landing_page_validator(self):
lp_validator = \
LandingPagesValidator(self.catalog, True)
with mock.patch(
'pydatajson.validators.url_validator.UrlValidator.is_working_url',
'pydatajson'
'.validators'
'.url_validator.UrlValidator.is_working_url',
return_value=(True, 200)):
res = lp_validator.validate()
self.assertEqual(0, len(list(res)))
Expand All @@ -105,7 +112,9 @@ def test_invalid_landing_page_validator(self):
lp_validator = \
LandingPagesValidator(self.catalog, True)
with mock.patch(
'pydatajson.validators.url_validator.UrlValidator.is_working_url',
'pydatajson'
'.validators'
'.url_validator.UrlValidator.is_working_url',
return_value=(False, 400)):
res = lp_validator.validate()
self.assertNotEqual(0, len(list(res)))
Expand All @@ -114,7 +123,9 @@ def test_valid_distribution_url_validator(self):
ddu_validator = \
DistributionUrlsValidator(self.catalog, True)
with mock.patch(
'pydatajson.validators.url_validator.UrlValidator.is_working_url',
'pydatajson'
'.validators'
'.url_validator.UrlValidator.is_working_url',
return_value=(True, 200)):
res = ddu_validator.validate()
self.assertEqual(0, len(list(res)))
Expand All @@ -123,7 +134,9 @@ def test_invalid_distribution_url_validator(self):
ddu_validator = \
DistributionUrlsValidator(self.catalog, True)
with mock.patch(
'pydatajson.validators.url_validator.UrlValidator.is_working_url',
'pydatajson'
'.validators'
'.url_validator.UrlValidator.is_working_url',
return_value=(False, 400)):
res = ddu_validator.validate()
self.assertNotEqual(0, len(list(res)))
Expand Down

0 comments on commit 5881dfc

Please sign in to comment.