Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Typo fixes #2054

Merged
merged 1 commit into from
Jun 13, 2016
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions docs/intro/tutorial.rst
Original file line number Diff line number Diff line change
Expand Up @@ -470,20 +470,20 @@ Here is a modification to our spider that does just that::
item['desc'] = sel.xpath('text()').extract()
yield item

Now the `parse()` method only extract the interesting links from the page,
Now the `parse()` method only extracts the interesting links from the page,
builds a full absolute URL using the `response.urljoin` method (since the links can
be relative) and yields new requests to be sent later, registering as callback
the method `parse_dir_contents()` that will ultimately scrape the data we want.

What you see here is the Scrapy's mechanism of following links: when you yield
What you see here is Scrapy's mechanism of following links: when you yield
a Request in a callback method, Scrapy will schedule that request to be sent
and register a callback method to be executed when that request finishes.

Using this, you can build complex crawlers that follow links according to rules
you define, and extract different kinds of data depending on the page it's
visiting.

A common pattern is a callback method that extract some items, looks for a link
A common pattern is a callback method that extracts some items, looks for a link
to follow to the next page and then yields a `Request` with the same callback
for it::

Expand Down