Skip to content
This repository has been archived by the owner on Feb 2, 2024. It is now read-only.

carsten-walther/crawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Simple CLI sitemap crawler

Crawler is a command line tool for crawling xml sitemaps, see https://www.sitemaps.org/.

Requirements

You have to install Go programming language. Also you have to install upx for packaging the binaries.

Compiling

Just run make all. All binaries were build into the build folder.

Installing

Once you build the binaries. Copy the desired binary depending on your operating system to /usr/local/bin/. You should rename it to a simpler filename like crawler. Also you can use the make install command.

Running

Crawler uses 4 parameters:

Parameter Type Description Default
url string the URL to your sitemap.xml
output string write results to a csv file
count int max count of URLs to crawl 0, means disabled
verbose bool show verbose output false

Run example:

$ crawler -url https://www.carstenwalther.de/sitemap.xml -count 100 -output results.csv -verbose

Console output:

Console

About

A simple CLI Sitemap Crawler

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published