Skip to content

Do different crawling projects share a browser pool? #2861

Answered by barjin
kanxue660 asked this question in Q&A

You must be logged in to vote

The current purpose of BrowserPool is to handle browser management during the crawler run and provide a unified interface for opening / closing pages in the managed browsers + handle fingerprint injection and proxy setup.

There indeed might be a performance hit from not reusing the managed browsers across multiple concurrent crawls. Unfortunately, right now, there is no way of instantiating the BrowserPool separately and passing it to the crawler instance. While there might be actual technical reasons for this (e.g. the way that proxies currently bind to running browsers), this is IMO rather a design oversight.

Currently, your best bets are:

  1. Run the projects separately and accept the per…

Replies: 2 comments 2 replies

You must be logged in to vote
1 reply
@B4nan

You must be logged in to vote
1 reply
@barjin

Answer selected by kanxue660
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
4 participants