Skip to content
#

scraping

Here are 2,862 public repositories matching this topic...

scrapy
Gallaecio
Gallaecio commented May 12, 2021

At the moment, ItemLoader(response=response) fails if response is not a TextResponse instance.

Passing a binary response can still be useful, though. For example, to allow processors to access the response from their loader context, and hence be able to report the source URL (response.url) when reporting input issues.

oldani
oldani commented Feb 18, 2019

If you're using proxies with requests-html and rendering JS sites is all good. Once you render a website pyppeteer don't know about this proxies and will expose your IP. This is an undesired behavior when scraping with proxies.

The idea is that whenever someone passes in proxies to the session object or any method call, make pyppeteer also use these proxies. #265

ferret
apify-js
mtrunkat
mtrunkat commented Sep 17, 2019

Main examples at Apify SDK webpage, Github repo and CLI templates should demonstrate how to manipulate with DOM and retrieve data from it.

Also add one example of scraping with Apify SDK + jQuery to https://sdk.apify.com/docs/examples/basiccrawler

Feedback from: https://medium.com/better-programming/do-i-need-python-scrapy-to-build-a-web-scraper-7cc7cac2081d

I lost an hour trying to make

jeverah
jeverah commented Feb 16, 2021

Hi,

I have read that it is possible to write extra parameters.
Where exactly do I do that?
Do I create a new file or can I add a line to my command in the terminal? Like --extra_info true?

### Optional parameters
*(For the 'get_posts' function)*.

- **group**: group id, to scrape groups instead of pages. Default is `None`.
- **pages**: how many pages of posts to request, usua

Improve this page

Add a description, image, and links to the scraping topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the scraping topic, visit your repo's landing page and select "manage topics."

Learn more