You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"...the dupe filter only catches requests after they leave the spider, so redirected pages are ignored by the dupe filter.
Since the dupefilter and the redirect middleware components are decoupled now, it would be awkward to implement what you suggest, but nevertheless I think it would be useful"
The text was updated successfully, but these errors were encountered:
Previously reported by michaelvmata on Trac http://dev.scrapy.org/ticket/299
If a crawl spider is redirected to an already visited page, it will still crawl it.
From the mailing list http://groups.google.com/group/scrapy-users/browse_thread/thread/ee9ad68f5dbacc6d:
"...the dupe filter only catches requests after they leave the spider, so redirected pages are ignored by the dupe filter.
Since the dupefilter and the redirect middleware components are decoupled now, it would be awkward to implement what you suggest, but nevertheless I think it would be useful"
The text was updated successfully, but these errors were encountered: