Seeking for dead links w/ DIY spider

WGET SPIDER

hard to read – where are broken links?

and not very effective (only first level?)

wget --spider -r --relative --no-directories --no-verbose https://nethence.com/
wget --spider -r --relative --no-directories --no-verbose https://pub.nethence.com/
#--level=99

# Total wall clock time: 54s

DIY single url

find dead links on a specific page — including those pointing to external sites

http://pub.nethence.com/bin/www/deadlinks.ksh.txt

http://pub.nethence.com/bin/www/deadlinks.bash.txt

deadlinks.ksh https://pub.nethence.com/xen/guest-sourcemage

Resources

How to check the first character in a string in Bash or UNIX shell? https://stackoverflow.com/questions/18488270/how-to-check-the-first-character-in-a-string-in-unix

How To Find Broken Links on Your Website Using Wget on Debian 7 https://www.digitalocean.com/community/tutorials/how-to-find-broken-links-on-your-website-using-wget-on-debian-7


HOME | GUIDES | LECTURES | LAB | SMTP HEALTH | HTML5 | CONTACT
Copyright © 2024 Pierre-Philipp Braun