Project

crawl

0.01
No commit activity in last 3 years
No release in over 3 years
Crawl all pages on a domain, checking for errors
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
 Dependencies
 Project Readme

Crawl

Circle CI

Crawl pages within a domain, reporting any page that returns a bad response code

Usage:

> crawl [options] domain

Usage: crawl [options] domain
-s, --start /home,/about         Starting path(s), defaults to /
-u, --username username          Basic auth username
-p, --password password          Basic auth password
-c, --connections count          Max mumber of parallel connections to use. The default is 5.
-v, --verbose                    Give details when crawling
-h, --help                       Show this message
    --version                    Print version

Example:

> crawl https://engineering.alphasights.com --connections=5 --start=/ --verbose

  Adding /
Fetching / ...
  Adding /positions/ruby-developer
  Adding /positions/js-ember-developer
  Adding /positions/ux-ui-designer
  Adding /positions/support-specialist
Fetching /positions/ruby-developer
Fetching /positions/js-ember-developer ...
Fetching /positions/ux-ui-designer ...
Fetching /positions/support-specialist ...

5 pages crawled without errors.

Copyright and License

Copyright AlphaSights and Contributors, 2015

MIT Licence