DPULSE - Domain Public Data Collection Service
-
Updated
Jun 4, 2024 - Python
DPULSE - Domain Public Data Collection Service
Weather data for Cardiff.
Apple's allowed autofill domains
API for voter registration status and sample ballots in Michigan.
An automated stock screening system which isolates top companies based on time-tested growth criteria.
The repository and website hosting the peer review process for new Programming Historian lessons
Obtaining ML and spreads for each MLB game querying an API
The Trends in Data Jobs project is a web scraping and data visualization tool designed to track and analyze trends in data-related job postings.
A Flask web application capable of scraping and parsing data from a single web page, manipulating the data within a Pandas DataFrame, and displaying the DataFrame on a webpage through multiple routes.
A bot that posts job openings at Reuters News
Net-Spider is a web scraping tool designed to retrieve the source code for a web page, including front-end elements such as JavaScript, CSS, images, and fonts. It allows you to crawl and download the source code from a target website.
Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.
procyclingstats scraper
Python toolkit for preprocessing data for the City Controller's Gun Violence Dashboard
Track changes on websites via git
Template of automated workflows using GitHub Actions with R code
Data analysis project to analyse the technologies requirements of the job market in Ile-de-France, France
A Python based web scraping api built with fastapi that provides easy access to manga contents
Web Scraping using Python Pandas and BeautifulSoup4
This project automates the scraping of news articles from the United Daily News (UDN) website, filters and processes them using specified keywords and OpenAI's GPT for Named Entity Recognition (NER), and exports the categorized data into a CSV file.
Add a description, image, and links to the web-scraping topic page so that developers can more easily learn about it.
To associate your repository with the web-scraping topic, visit your repo's landing page and select "manage topics."