Fetching contributors…
Cannot retrieve contributors at this time
161 lines (112 sloc) 5.79 KB

Cluster setup guide

This guide is targeting an initial setup of crawling cluster, probably further tuning will be needed. This guide implies you use Kafka message bus for cluster setup (recommended), although it is also possible to use ZeroMQ, which is less reliable option.

Things to decide

  • The speed you want to crawl with,
  • number of spider processes (assuming that single spider process gives a maximum of 1200 pages/min),
  • number of DB and Strategy worker processes.

Things to setup before you start

Things to implement before you start

Configuring Kafka

Create all topics needed for Kafka message bus

Configuring HBase

Configuring Frontera

Every Frontera component requires its own configuration module, but some options are shared, so we recommend to create a common modules and import settings from it in component's modules.

  1. Create a common module and add there:

    from __future__ import absolute_import
    from frontera.settings.default_settings import MIDDLEWARES
    SPIDER_FEED_PARTITIONS = 2 # number of spider processes
    SPIDER_LOG_PARTITIONS = 2 # worker instances
    KAFKA_LOCATION = 'localhost:9092' # your Kafka broker host:port
    SCORING_TOPIC = 'frontier-scoring'
  2. Create workers shared module:

    from __future__ import absolute_import
    from .common import *
    BACKEND = 'frontera.contrib.backends.hbase.HBaseBackend'
    HBASE_THRIFT_HOST = 'localhost' # HBase Thrift server host and port
  3. Create DB worker module:

    from __future__ import absolute_import
    from .worker import *
    LOGGING_CONFIG='logging-db.conf' # if needed
  4. Create Strategy worker's module:

    from __future__ import absolute_import
    from .worker import *
    CRAWLING_STRATEGY = '' # path to the crawling strategy class
    LOGGING_CONFIG='logging-sw.conf' # if needed

The logging can be configured according to see the :doc:`list of loggers <loggers>`.

  1. Configure spiders module:

    from __future__ import absolute_import
    from .common import *
    BACKEND = 'frontera.contrib.backends.remote.messagebus.MessageBusBackend'
    LOCAL_MODE = False  # by default Frontera is prepared for single process mode

6. Configure Scrapy settings module. It's located in Scrapy project folder and referenced in scrapy.cfg. Let's add there:

FRONTERA_SETTINGS = ''  # module path to your Frontera spider config module


    'frontera.contrib.scrapy.middlewares.schedulers.SchedulerSpiderMiddleware': 999,
    'frontera.contrib.scrapy.middlewares.seeds.file.FileSeedLoader': 1,
    'frontera.contrib.scrapy.middlewares.schedulers.SchedulerDownloaderMiddleware': 999,

Starting the cluster

First, let's start storage worker:

# start DB worker only for batch generation
# use single instance for every 10 partitions
$ python -m frontera.worker.db --config [db worker config module] --no-incoming --partitions 0,1

# Optionally, start next one dedicated to spider log processing.
$ python -m frontera.worker.db --no-batches --config [db worker config module]

Next, let's start strategy workers, one process per spider log partition:

$ python -m frontera.worker.strategy --config [strategy worker config] --partition-id 0
$ python -m frontera.worker.strategy --config [strategy worker config] --partition-id 1
$ python -m frontera.worker.strategy --config [strategy worker config] --partition-id N

You should notice that all processes are writing messages to the log. It's ok if nothing is written in streams, because of absence of seed URLs in the system.

Let's put our seeds in text file, one URL per line and run:

$ python -m frontera.utils.add_seeds --config [your_frontera_config] --seeds-file [path to your seeds file]

Finally, a single spider per spider feed partition:

$ scrapy crawl [spider] -L INFO -s SPIDER_PARTITION_ID=1
$ scrapy crawl [spider] -L INFO -s SPIDER_PARTITION_ID=2
$ scrapy crawl [spider] -L INFO -s SPIDER_PARTITION_ID=N

You should end up with N spider processes running. Also :setting:`SPIDER_PARTITION_ID` can be read from config file.

You're done, crawler should start crawling. Any component can be restarted any time, without major data loss. However, for pausing its enough to stop batch gen only.