From f83b2d1be4908b3e8e2efbe9fbeb219c3afde2bd Mon Sep 17 00:00:00 2001 From: "deepsource-autofix[bot]" <62050782+deepsource-autofix[bot]@users.noreply.github.com> Date: Sun, 4 Oct 2020 04:59:27 +0000 Subject: [PATCH] Change methods not using its bound instance to staticmethods --- Scripts/Bots/Twitter_bot/Twitter_bot.py | 3 ++- .../IMDB-Mass-Scraper/middlewares.py | 21 ++++++++++++------- 2 files changed, 16 insertions(+), 8 deletions(-) diff --git a/Scripts/Bots/Twitter_bot/Twitter_bot.py b/Scripts/Bots/Twitter_bot/Twitter_bot.py index 512d59b22..ae8fdc2e8 100644 --- a/Scripts/Bots/Twitter_bot/Twitter_bot.py +++ b/Scripts/Bots/Twitter_bot/Twitter_bot.py @@ -2,7 +2,8 @@ from twython import TwythonStreamer import configparser class MyStreamer(TwythonStreamer): - def on_success(self, data): + @staticmethod + def on_success(data): if 'text' in data: a=data['text'].lower() username=data['user']['screen_name'] diff --git a/Scripts/Web_Scrappers/IMDB-Mass-Scraper/middlewares.py b/Scripts/Web_Scrappers/IMDB-Mass-Scraper/middlewares.py index 921ddd670..b43e560bb 100644 --- a/Scripts/Web_Scrappers/IMDB-Mass-Scraper/middlewares.py +++ b/Scripts/Web_Scrappers/IMDB-Mass-Scraper/middlewares.py @@ -21,14 +21,16 @@ def from_crawler(cls, crawler): crawler.signals.connect(s.spider_opened, signal=signals.spider_opened) return s - def process_spider_input(self, response, spider): + @staticmethod + def process_spider_input(response, spider): # Called for each response that goes through the spider # middleware and into the spider. # Should return None or raise an exception. return None - def process_spider_output(self, response, result, spider): + @staticmethod + def process_spider_output(response, result, spider): # Called with the results returned from the Spider, after # it has processed the response. @@ -43,7 +45,8 @@ def process_spider_exception(self, response, exception, spider): # Should return either None or an iterable of Request or item objects. pass - def process_start_requests(self, start_requests, spider): + @staticmethod + def process_start_requests(start_requests, spider): # Called with the start requests of the spider, and works # similarly to the process_spider_output() method, except # that it doesn’t have a response associated. @@ -52,7 +55,8 @@ def process_start_requests(self, start_requests, spider): for r in start_requests: yield r - def spider_opened(self, spider): + @staticmethod + def spider_opened(spider): spider.logger.info('Spider opened: %s' % spider.name) @@ -68,7 +72,8 @@ def from_crawler(cls, crawler): crawler.signals.connect(s.spider_opened, signal=signals.spider_opened) return s - def process_request(self, request, spider): + @staticmethod + def process_request(request, spider): # Called for each request that goes through the downloader # middleware. @@ -80,7 +85,8 @@ def process_request(self, request, spider): # installed downloader middleware will be called return None - def process_response(self, request, response, spider): + @staticmethod + def process_response(request, response, spider): # Called with the response returned from the downloader. # Must either; @@ -99,5 +105,6 @@ def process_exception(self, request, exception, spider): # - return a Request object: stops process_exception() chain pass - def spider_opened(self, spider): + @staticmethod + def spider_opened(spider): spider.logger.info('Spider opened: %s' % spider.name)