Skip to content

olshevskiy87/urlook

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

67 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

URLook

Build Status MIT license Go Report Card

Find and check http(s) links in the text file.

Install

$ go get github.com/olshevskiy87/urlook

Usage

Usage: urlook [--fail-on-duplicate] [--timeout TIMEOUT] [--white WHITE] [FILENAMES [FILENAMES ...]]

Positional arguments:
  FILENAMES              filenames with links to check

Options:
  --fail-on-duplicate    fail if there is a duplicate url
  --timeout TIMEOUT, -t TIMEOUT
                         request timeout in seconds [default: 10]
  --white WHITE, -w WHITE
                         white list url (can be specified multiple times)
  --help, -h             display this help and exit
  --version              display version and exit

Examples

$ urlook test_links
URLs to check: 7
  1. http://cb.vu/unixtoolbox.xhtml
  2. https://eax.me
  3. http://eax.me
  4. https://google.com
  5. https://github.com
  6. https://github.com/non-existent-url-non-existent-url-non-existent-url
  7. https://www.reddit.com
✓→✓→✓x
  1. http://eax.me [301, Moved Permanently] -> https://eax.me/
  2. https://google.com [302, Found] -> https://www.google.ru/?gfe_rd=cr&dcr=0&ei=EFSZWqWQAcGDtAGArrLYBw
  3. https://github.com/non-existent-url-non-existent-url-non-existent-url [404, Not Found]
issues found: 3
$ grep github test_links | urlook --white non-existent-url
URLs to check: 1
  1. https://github.com
✓
White listed URLs (1):
 - https://github.com/non-existent-url-non-existent-url-non-existent-url

no issues found
$ echo 'check these links: https://ya.ru, https://www.reddit.com' | urlook
URLs to check: 2
  1. https://ya.ru
  2. https://www.reddit.com
✓✓
no issues found
$ echo 'https://ya.ru https://github.com https://ya.ru' | urlook --fail-on-duplicate
URLs to check: 2
  1. https://github.com
  2. https://ya.ru
✓✓
Duplicates:
 - https://ya.ru (2)
duplicates found: 1

Todo

  • read input from stdin
  • remove duplicate urls
  • add cli-option to fail on duplicate url
  • add cli-option timeout
  • add white list option
  • read more than one file
  • add tests
  • add option to allow redirects

Motivations

Inspired by ruby gem awesome_bot

License

MIT. See LICENSE for details.

About

Find and check http(s) links in the text file

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published