Bristol SEO Meetup: Technical SEO Audit For Beginners
At the November 2018 Bristol SEO Meetup, I shared a presentation on technical SEO audits for beginners, and you can check out the slides here.
In this talk, I covered:
- SEO Tools
- HTTP & HTTPS
- 404 Pages
- Redirect Errors
- Page Speed
- Structured data
Some of the tools that are referenced throughout this presentation, and that we use in our day to day activities include:
ScreamingFrog: ScreamingFrog is one of our favourite SEO tools and we use it every day to crawl websites and identify any SEO issues, such as 404 errors, incorrect canonical URLs, and redirect chains, and to also extract key on-page SEO elements, such as title tags, meta descriptions, and structured data.
Ahrefs: We use Ahrefs to mainly get an in-depth look into a website’s backlink profile. We are able to monitor and identify new or lost backlinks, and research competitors’ backlink profiles to get an idea of the quality of their links and to identify opportunities for new links.
SEMrush: SEMrush is our go-to tool for tracking keyword movements, competitor performance, and for running daily, weekly, or monthly site audits that will quickly flag any new SEO issues so we can act immediately to minimise impact.
GTMetrix: We use GTMetrix to test a website’s page speed. There are so many tools that you can use to review page speed, such as Google’s PageSpeed Insights, but we like to use GTMetrix as it gives a much more in-depth breakdown of which elements are slowing down your webpage and why.
TinyPNG: TinyPNG is a tool you can use to optimise your images for page speed. By uploading your image files to TinyPNG, you can reduce the file size of your images which helps reduce the page load time.
CSS Minifier: CSS Minifer is also related to page speed and is used to minify CSS files to reduce their file size, but make sure you speak to your development team before attempting to implement this!
So, let’s start at the very beginning, the robots.txt file. Your robots.txt file can usually be found by adding /robots.txt at the end of any root domain, e.g. https://example.com/robots.txt.
The robots.txt file instructs web robots, usually search engines robots such as Googlebot, if they can or cannot crawl specific parts of your website. When a search bot reaches a website, they will first look for the robots.txt file and read this before continuing to crawl the site. By using the disallow directive, you can block search engines from crawling specific pages or sections of your website, such as your staging environment.
Examples of robots.txt:
Blocking all web crawlers from crawling your entire website:
Blocking Googlebot from crawling a specific URL and its sub-folders:
Once your website has been crawled, it can then be indexed in the search results.
It’s important to review how your website appears in the search index, and to identify:
To find out which of your website’s pages are being included in the index, you can perform a site:search in Google’s search bar e.g. site:edicomedia.com, and this returns a list of pages that are in Google’s index.
There is also a new tool in Google Search Console, the Index Coverage Report, that allows you to find any URLs that are included in the index and identify any that are excluded and why they are being excluded.
If your webpages are being excluded from the index:
Your website’s sitemap can usually be found by adding /sitemap.xml or /sitemap at the end of your root domain e.g. example.com/sitemap.xml. Your sitemap helps search engines better understand your website so that they crawl it more intelligently.
Using a XML to CSV converter, you can export your website’s sitemap to a spreadsheet and upload the URLs to ScreamingFrog and:
URL Status Codes
Put together a list of all of your website’s URLs and identify any that do not return a 200 status code. We use several tools to collect URLs and check their HTTP status, including ScreamingFrog, Redirect Path, Search Console, and Ahrefs.
Common HTTP Status Codes:
HTTP Status Code 3xx – Redirection
HTTP Status Code 4xx – Client Error
Perform a crawl of your website to identify any internal or external links that result in a 4xx or 3xx status code. If an internally linked URL returns a 3xx status code, you should update the link to the destination URL. If it returns a 4xx error, you should attempt to find a relevant webpage that it can then be redirected to, however if there are no relevant destinations, remove the internal link. Also do this for any external links you have on your website.
Export and crawl your backlink profile to identity any inbound links to URLs that return a 4xx status code:
404 Page Not Found
When a user does land on a URL that returns a 404 error, they should land on a page that is designed in a way that allows the user to easily navigate your website so that they don’t immediately leave. On your 404 page, you should:
One of the most common mistakes we see with a 404 page is that the URL actually redirects to a /page-not-found URL which returns a 200 status code. A broken URL should always return a 404 error.
By now, your website should really be HTTPS. Since 2014, Google has been using HTTPS as a ranking signal and now over 70% of Google’s page on search results are HTTPS.
Google Chrome now also marks all non-HTTPS URLs as non-secure and since October 2018, this will turn red when a user enters sensitive data.
You should ensure that all of your URLs are HTTPS and your HTTP URLs correctly 301 redirect to the HTTPS version to avoid creating duplicate URLs and content.
There are several ways to write a URL and while no style is officially preferred by Google from an SEO perspective, you should always be consistent in your choice and ensure that one version correctly 301 redirects to your preferred version.
So, if you don’t use www in your URLs, you should make sure that the www version correctly redirects to the non-www version, or if you add a trailing slash to the end of your URLs, make sure that the URL without a trailing slash redirects correctly to the trailing slash version and vice versa. Otherwise you will be creating two completely different URLs that have the same content.
There are multiple tools you can use to review your page speed, including:
We prefer to use GTMetrix as it provides a more indepth analysis of your page speed and what can be done to improve it.
Some of the most common fixes we come across when analysing page speed are:
The file size of your images can have a significant impact on the page load time of your webpage, and there are several tools that you can use to compress your images and reduce their file size, such as TinyPNG.
Structured data is becoming increasingly more common and more options and opportunities are being created all the time. It helps search engines better understand the content on a webpage and can enhance how your content or listing appears in the search results.
Structured data can be added to several elements of a website, such as:
You should review all the content types on your website and identify any structured data opportunities.
In April 2015, the Mobile-Friendly algorithm was released and was designed to give mobile-friendly pages a boost in mobile search results, and this was updated again in 2016. In May 2018, Google rolled out their mobile-first index, which means your website will now be ranked based on the mobile-friendly version of your website, even for search results shown to desktop users. So, it’s super important that your website is fully responsive. Read our blog post on the Mobile-First Index to find out more about what you can do to make sure your website is ready.
If you are interested in attending similar talks, the Bristol SEO meetup is a monthly event that is free for all to attend, you can sign up to the event and get email reminders for upcoming events.