Skip to content

Commit

Permalink
Run autopep8 --max-line-length 119 --in-place -r .
Browse files Browse the repository at this point in the history
  • Loading branch information
jpmckinney committed May 27, 2020
1 parent 9de01e7 commit 958370b
Show file tree
Hide file tree
Showing 18 changed files with 26 additions and 27 deletions.
2 changes: 1 addition & 1 deletion kingfisher_scrapy/base_spider.py
Expand Up @@ -214,7 +214,7 @@ def parse_zipfile(self, response, data_type, file_format=None, encoding='utf-8')
encoding=encoding, file_name=filename)
else:
yield self.build_file(data.read(), filename, data_type=data_type, url=response.request.url,
encoding=encoding)
encoding=encoding)
else:
yield self.build_file_error_from_response(response)

Expand Down
1 change: 0 additions & 1 deletion kingfisher_scrapy/items.py
Expand Up @@ -14,7 +14,6 @@ class File(scrapy.Item):
files_store = scrapy.Field()



class FileItem(scrapy.Item):
number = scrapy.Field()
file_name = scrapy.Field()
Expand Down
10 changes: 5 additions & 5 deletions kingfisher_scrapy/spiders/afghanistan_records.py
Expand Up @@ -46,10 +46,10 @@ def parse_record(self, response):
url = response.request.url
# This is dangerous as we might get stuck in a loop here if we always get a 429 response. Try this for now.
yield scrapy.Request(
url=url,
meta={'kf_filename': url.split('/')[-1]+'.json'},
callback=self.parse_record,
dont_filter=True,
)
url=url,
meta={'kf_filename': url.split('/')[-1]+'.json'},
callback=self.parse_record,
dont_filter=True,
)
else:
yield self.build_file_error_from_response(response)
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/armenia.py
Expand Up @@ -20,7 +20,7 @@ def parse(self, response):
if response.status == 200:

yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')

json_data = json.loads(response.text)
if not (self.sample):
Expand Down
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/australia_nsw.py
Expand Up @@ -67,7 +67,7 @@ def parse_list(self, response):
def parse(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')

else:
yield self.build_file_error_from_response(response)
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/canada_buyandsell.py
Expand Up @@ -30,6 +30,6 @@ def start_requests(self):
def parse(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')
else:
yield self.build_file_error_from_response(response)
8 changes: 4 additions & 4 deletions kingfisher_scrapy/spiders/chile_base.py
Expand Up @@ -27,9 +27,9 @@ def get_year_month_until(self):

def get_sample_request(self):
return scrapy.Request(
url=self.base_list_url.format(2017, 10, 0, 10),
meta={'year': 2017, 'month': 10}
)
url=self.base_list_url.format(2017, 10, 0, 10),
meta={'year': 2017, 'month': 10}
)

def start_requests(self):
if self.sample:
Expand Down Expand Up @@ -85,4 +85,4 @@ def base_parse(self, response, package_type):
return [self.build_file_error_from_response(response, errors={'http_code': data['status']})]
else:
return [self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='%s_package' % package_type)]
data_type='%s_package' % package_type)]
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/colombia.py
Expand Up @@ -45,7 +45,7 @@ def parse(self, response):
elif response.status == 200:

yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')

if not self.sample:
yield self.next_link(response)
Expand Down
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/dominican_republic.py
Expand Up @@ -38,7 +38,7 @@ def parse(self, response):
for f in tmpfile.infolist():
with tmpfile.open(f) as jsonFile:
yield self.build_file(jsonFile.read(), f.filename, data_type='release_package',
url=response.request.url)
url=response.request.url)
os.remove(file.name)
else:
filename = response.request.url.split('/')[-1]
Expand Down
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/mexico_cdmx.py
Expand Up @@ -34,6 +34,6 @@ def parse_list(self, response):
def parse_record(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')
else:
yield self.build_file_error_from_response(response)
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/mexico_grupo_aeroporto.py
Expand Up @@ -15,7 +15,7 @@ def start_requests(self):
def parse(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')

else:
yield self.build_file_error_from_response(response)
4 changes: 2 additions & 2 deletions kingfisher_scrapy/spiders/mexico_jalisco.py
Expand Up @@ -41,13 +41,13 @@ def parse_record_package(self, response):
callback=self.parse_release_package
)
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='record_package')
data_type='record_package')
else:
yield self.build_file_error_from_response(response)

def parse_release_package(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')
else:
yield self.build_file_error_from_response(response)
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/moldova.py
Expand Up @@ -27,7 +27,7 @@ def parse(self, response):
if response.status == 200:
if response.request.meta['data']:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='record_package')
data_type='record_package')
else:
self.build_file_from_response(response, response.request.meta['kf_filename'])
json_data = json.loads(response.text)
Expand Down
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/moldova_old.py
Expand Up @@ -22,6 +22,6 @@ def start_requests(self):
def parse(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')
else:
yield self.build_file_error_from_response(response)
4 changes: 2 additions & 2 deletions kingfisher_scrapy/spiders/paraguay_hacienda.py
Expand Up @@ -21,7 +21,7 @@ class ParaguayHacienda(BaseSpider):

custom_settings = {
'DOWNLOADER_MIDDLEWARES': {
'kingfisher_scrapy.middlewares.ParaguayAuthMiddleware': 543,
'kingfisher_scrapy.middlewares.ParaguayAuthMiddleware': 543,
},
'CONCURRENT_REQUESTS': 1,
}
Expand Down Expand Up @@ -81,7 +81,7 @@ def parse(self, response):
)
else:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')

else:
yield self.build_file_error_from_response(response)
Expand Down
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/scotland.py
Expand Up @@ -54,6 +54,6 @@ def start_requests(self):
def parse(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')
else:
yield self.build_file_error_from_response(response)
2 changes: 1 addition & 1 deletion kingfisher_scrapy/spiders/test_fail.py
Expand Up @@ -35,7 +35,7 @@ def start_requests(self):
def parse(self, response):
if response.status == 200:
yield self.build_file_from_response(response, response.request.meta['kf_filename'],
data_type='release_package')
data_type='release_package')

else:

Expand Down
2 changes: 1 addition & 1 deletion tests/test_extensions.py
Expand Up @@ -70,7 +70,7 @@ def test_item_scraped_file(sample, is_sample, path, note, encoding, encoding2, d
if encoding:
kwargs['encoding'] = encoding
item = spider.build_file(b'{"key": "value"}', 'file.json', url='https://example.com/remote.json',
data_type='release_package', post_to_api=post_to_api, **kwargs)
data_type='release_package', post_to_api=post_to_api, **kwargs)

store_extension.item_scraped(item, spider)

Expand Down

0 comments on commit 958370b

Please sign in to comment.