Tutorial

How To Find Broken Links on Your Website Using Wget on Debian 7

Published on April 14, 2015
How To Find Broken Links on Your Website Using Wget on Debian 7

Introduction

How many times have you clicked a HTML link on a webpage only to get a 404 Not Found error? Broken links exist because webpages sometimes get moved or deleted over time. It is the job of the webmaster to find those broken links before the human web visitors or the search engine robots do. Delay in correcting the problem results in bad user experience and possible penalty for search engine page ranking.

If your website contains more than a few pages, manually checking each individual link becomes too labor intensive, but there are numerous tools that automate that task. You can use web-based apps, like ones provided by Google Webmaster Tools and the World Wide Web Consortium (W3C), but they generally lack more advanced features. If you run WordPress, you could use a plugin, but some shared web hosting companies ban them because they run on the same server as the website, and link checking is resource-intensive.

Another option is to use a Linux-based program on a separate machine. These include general web crawlers that also uncover broken links (like wget) and custom-built link checkers (like linkchecker and klinkstatus). They are highly customizable and minimize any negative impact on the response time of your target website.

This tutorial explains how to use wget to find all of the broken links on a website so you can correct them.

Prerequisites

To follow this tutorial, you will need:

  • Two Debian 7 Droplets, one generic machine to run wget from (generic-1) and one which hosts your website (webserver-1).

  • A sudo non-root user on both generic-1 and webserver-1. Click here for instructions.

  • webserver-1 needs to have the LAMP stack installed. Click here for instructions.

  • Optionally, the webserver can have its own registered domain name. If so, use your domain name wherever you see your_server_ip. Click here for instructions.

Although this tutorial is written for Debian 7, the wget examples should also run on other modern Linux distributions. You may need to install wget on other distributions where it is not included by default.

Step 1 — Creating a Sample Webpage

First, we’ll add a sample webpage with multiple missing links.

Log into webserver-1. Open a new file called spiderdemo.html for editing using nano or your favorite text editor.

sudo nano /var/www/spiderdemo.html

Paste the following into the file. This is a very simple webpage that includes two broken links, one internal (add in your server IP where it’s highlighted below) and one external.

<html>
<head> <title>Hello World!</title> </head>
<body>

<p>
<a href="http://your_server_ip/badlink1">Internal missing link</a>.
<a href="https://www.digitalocean.com/thisdoesntexist">External missing link</a>.
</p>

</body>
</html>

Save and close the file.

Next, change the file owner and group of spiderdemo.html to the default webserver user, www-data.

sudo chown www-data:www-data /var/www/spiderdemo.html

Finally, change the file permissions of the new HTML file.

sudo chmod 664  /var/www/spiderdemo.html

You can now see the example page at http://your_server_ip/spiderdemo.html.

Step 2 — Running wget

wget is a general-purpose website downloader which can also be used as a web crawler. In this step, we’ll configure wget to report whether each link points to an existing page or is broken without downloading the page.

Note: Only check links on a website which you own. Link checking on a website incurs significant computing overhead, so these activities may be interpreted as spamming.

Log into generic-1 and run the following wget command. Explanations of each flag are below; you can modify this command for your use case.

wget --spider -r -nd -nv -H -l 1 -w 2 -o run1.log  http://your_server_ip/spiderdemo.html

The following are the basic flags you’ll need:

  • --spider stops wget from downloading the page.

  • -r makes wget recursively follow each link on the page.

  • -nd, short for --no-directories, prevents wget from creating a hierarchy of directories on your server (even when it is configured to spider only).

  • -nv, short for --no-verbose, stops wget from outputting extra information that is unnecessary for identifying broken links.

The following are optional parameters which you can use to customize your search:

  • -H, short for --span-hosts,makes wget crawl to subdomains and domains other than the primary one (i.e. external sites).

  • -l 1 is short for --level. By default, wget crawls up to five levels deep from the initial URL, but here we set it to one. You may need to play with this parameter depending on the organization of your website.

  • -w 2, short for --wait, instructs wget to wait 2 seconds between requests to avoid bombarding the server, minimizing any performance impact.

  • -o run1.log saves wget’s output to a file called run1.log instead of displaying it in your terminal.

After you run the above wget command, extract the broken links from the output file using the following command.

grep -B1 'broken link!' run1.log

The -B1 parameter specifies that, for every matching line, wget displays one additional line of leading context before the matching line. This preceding line contains the URL of the broken link. Below is sample output from the above grep command.

http://your_server_ip/badlink1:
Remote file does not exist -- broken link!!!
https://www.digitalocean.com/thisdoesntexist:
Remote file does not exist -- broken link!!!

Step 3 — Finding Referrer URLs

Step 2 reports the broken links but doesn’t identify the referrer webpages, i.e., the pages on your site that contain those links. In this step, we’ll find the referrer webpages.

A convenient way to identify the referrer URL is by examining the webserver’s access log. Log in to webserver-1 and search the Apache logs for the broken link.

sudo grep Wget /var/log/apache2/access.log | grep "HEAD /badlink1"    

The first grep in the above command finds all access requests by wget to the webserver. Each access request includes the User Agent string, which identifies the software agent responsible for generating the web request. The User Agent*identifier for wget is Wget/1.13.4 (linux-gnu).

The second grep searches for the partial URL of the broken link (/badlink1). The partial URL used is that part of the URL that follows the domain.

Sample output from the grep command chain is as follows:

111.111.111.111 - - [10/Apr/2015:17:26:12 -0800] "HEAD /badlink1 HTTP/1.1" 404 417 "http://your_server_ip/spiderdemo.html" "Wget/1.13.4 (linux-gnu)"   

The referrer URL is the second last item on the line: http://your_server_ip/spiderdemo.html.

Conclusion

This tutorial explains how to use the wget tool to find the broken links on a website, and how to find the referrer pages which contain those links. You can now make corrections by updating or removing any broken links.

Thanks for learning with the DigitalOcean Community. Check out our offerings for compute, storage, networking, and managed databases.

Learn more about us


About the authors

Default avatar

staff technical writer

hi! i write do.co/docs now, but i used to be the senior tech editor publishing tutorials here in the community.


Still looking for an answer?

Ask a questionSearch for more help

Was this helpful?
 
3 Comments


This textbox defaults to using Markdown to format your answer.

You can type !ref in this text area to quickly search our full set of tutorials, documentation & marketplace offerings and insert the link!

This would have been useful if it simply scanned an arbitrary web page (provided as an input), and reported all broken links it found. You could have made this simple, yet still cover the ‘droplets’, dual machines & lecture on bandwidth conservation. Sheesh.

Checking a site for broken links is always a great idea. It improves usability and is a strong indicator that you are actively improving your site…or at least keeping an eye on it, which search engines like to see. So there are obvious SEO benefits to making sure your site doesn’t have any broken links. Although Wget seems like a pretty good tool, I recommend that my clients who are less tech savvy use a free broken link check tool to locate broken links. There are many good options online that allow you to simply enter your domain, without having to sign up or register for a service. Then you can locate and fix links any time. Although the frequency at which one should check will depend on how many links you have throughout your site, it’s at least a good thing to do once a month.

Try DigitalOcean for free

Click below to sign up and get $200 of credit to try our products over 60 days!

Sign up

Join the Tech Talk
Success! Thank you! Please check your email for further details.

Please complete your information!

Get our biweekly newsletter

Sign up for Infrastructure as a Newsletter.

Hollie's Hub for Good

Working on improving health and education, reducing inequality, and spurring economic growth? We'd like to help.

Become a contributor

Get paid to write technical tutorials and select a tech-focused charity to receive a matching donation.

Welcome to the developer cloud

DigitalOcean makes it simple to launch in the cloud and scale up as you grow — whether you're running one virtual machine or ten thousand.

Learn more
DigitalOcean Cloud Control Panel