-
Notifications
You must be signed in to change notification settings - Fork 405
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Passes meta information from previous request #145
Comments
I'm not sure if this works like this because of the asynchronicity, but everytime |
I noticed it as well, and my best guess so fa with this is that we could store this lis on a global variable, because the order is correct and then on the preRequest match the future request with this global variable. But I am also thinking that this option could also be useful for example configure the referrer for the next request. As far I understand, currently all the requests won't have any referrer and this can set off a few alarms and got blocked. |
@yvmarques |
@yujiosaka I am not sure in the The idea would to have something similar to what Scrapy has for the https://doc.scrapy.org/en/latest/topics/request-response.html#scrapy.http.Request.meta |
Yes, it will be. I just thought you only wanted to know where the request is coming from. If it's what you wanted, I can probably add the feature quick. |
I don't know how hard would it be to, for example get the result of a previous request passed to |
I am not sure how I can archive this, but my requirement it's that I need to know the order of the links of the initial request and pass it to the next request in order to save it with more data from the link's request.
Let's say that my initial request contains three pages in that order:
foo.html -> 1st link in the HTML
bar.html -> 2nd link in the HTML
baz.html -> etc.
When I will request foo.html (because I configured the crawler to depth: 2) I would like to know that this page was the 1st link from the previous page.
Is that possible ?
Thanks,
The text was updated successfully, but these errors were encountered: