Asynchronous web crawling and scraping with Python for minimalists
- Crawling: download an entire website in just a few seconds
- Scraping: Customizable XPath & CSS data selectors
- Zero-configuration: get up and running with ~5 LoC
- Interfaces: Python script + Web UI + JSON API
Built with:
- `asyncio`
- `aiohttp`
- `parsel`
- `FastAPI`
- `VueJS`
pip install crawlio
Create a custom Crawler
instance and run it
import asyncio
from crawlio import Crawler, Selector
bot = Crawler(
url='https://quotes.toscrape.com/',
selectors=[
Selector('links', '//a/@href'),
Selector('heading', type='xpath', query='//h3//text()', process=lambda items: ' '.join(items))
]
)
output = asyncio.run(bot.run())
for item in output["data"]:
print(item)
Copyright (C) 2022-2023 Maximilian Wolf
This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version.
This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
You should have received a copy of the GNU General Public License along with this program. If not, see https://www.gnu.org/licenses/.