This repo provide its users with a script that enables them to crawl and take screenshots of as many pages as they want to.
- Upload the list of URLs you want to parse in the
pages.json
file - Run
yarn start
ornpm start
- Once complete, your screenshots will be available within a "screenshots" folder
- You can change the parallelism parameter in
crawler.js
. This will enable your machine to load more tabs at once but will consume more ressources.
Note: if you are looking for a tool that will crawl a site on its own, you may be better off using Pappet.