Skip to main content

MongoDB-based components for Scrapy

Project description

# Scrapy MongoDB Queue MongoDB-based components for scrapy that allows distributed crawling

# Available Scrapy components: * Scheduler * Duplication Filter

Installation

From pypi:

$ pip install scrapy-mongodb-queue

From github:

$ git clone https://github.com/jbinfo/scrapy-mongodb-queue.git
$ cd scrapy-mongodb-queue
$ python setup.py install

Usage

Enable the components in your settings.py:

# Enables scheduling storing requests queue in redis.
SCHEDULER = "scrapy_mongodb_queue.scheduler.Scheduler"

# Don't cleanup mongodb queues, allows to pause/resume crawls.
MONGODB_QUEUE_PERSIST = True

# Specify the host and port to use when connecting to Redis (optional).
MONGODB_SERVER = 'localhost'
MONGODB_PORT = 27017
MONGODB_DB = "my_db"

# MongoDB collection name
MONGODB_QUEUE_NAME = "my_queue"

Author

This project is maintained by Lhassan Baazzi ([GitHub](https://github.com/jbinfo) | [Twitter](https://twitter.com/baazzilhassan) | [LinkedIn](https://ma.linkedin.com/pub/lhassan-baazzi/49/606/a70))

Project details


Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page