Skip to content

danielMensah/webcrawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Web Crawler

This is a simple webcrawler that crawls a given URL as well as visiting the links on the page. This crawler avoids visiting the same URL twice, and any external links are ignored.

Requirements

These dependencies can also be installed with Homebrew.

  • Requires Go 1.18 or greater. This can be installed with brew brew install go or downloaded here.
  • Requires Golangci Lint. This can be installed with brew brew install golangci-lint or downloaded here.

Install Dependencies

Install dependencies, issue the following command(s):

make install

Testing and Formatting

To run the tests, issue the following command(s):

make test

Lint only

Run linting only:

make lint

How to Run

To run the application with the default settings, simply issue the following example command(s):

make run https://example.com/

If you wish to change the default settings, navigate to cmd/main.go and change the default settings. Here is the list of settings that can be changed:

  • retryMax: The maximum number of times to retry a failed request.
  • retryMaxWait: The maximum amount of time to wait before retrying a failed request.
  • workers: The number of concurrent workers to use.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published