Skip to content

dshah3/Coronavirus-US-Case-Updates

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 

Repository files navigation

Coronavirus US Case Updates

Scrapy Web Scraper with Slack Webhooks to Periodically Send US Coronavirus Case Updates to a Desired Slack Channel

Table of Contents

  • Installation of Necessary Modules and Packages
  • Slack Integration
  • deferLater and sleep Capabilities
  • Resources

Installation of Modules and Packages
For webscrapers, I recommend using Scrapy: it has great documentation, fairly easy to use, and supports lots of integrations to 3rd party software. However, it does not support periodic jobs, so you can integrate Scrapy with Scrapy-Do or the deferLater function. You can use pip to install Scrapy and the integrated terminal in your IDE to check the version of Scrapy.

Slack Integration
We can integrate Slack into our project using the Slack API and Slack Webhooks. Create a Slack app and webhook for the desired channel, and authorize your app to have access to a certain channel. Insert your webhook url into the proper section of the project, and use a cURL command to send a message to slack. You can convert a cURL command to python using this website.

deferLater and sleep Capabilities
Since we are running Scrapy through a script, we must use a Twisted Crawler and Twisted deferLater in order to start and stop our process, and also run the spider periodically. I use a Callback function to add a delay through a deferLater function..

Resources - Scrapy Documentation - Scrapy-Do - Slack Webhooks - Crawler Process - cURL Command Converter

About

Python Web Scraper with Slack Message Integration

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages