• SEO Tips

7 Technical SEO Issues You Forgot to Check

  • David Ryan
  • 3 min read
7 Technical SEO Issues You Forgot to Check


The life of a technical SEO can be very fast-paced. A pagespeed audit one week, international considerations the next, content pruning after that, etc. etc.

Technical SEOs can get so lost in a meaty (or vegan) task, that they often overlook the smaller fixes which can have a big impact on your organic performance.

These aren’t technical SEO silver bullets, these are recommendations that you know you should fix but never do.

Therefore, I’m going to share some of my favourite “oh yeah, forgot to check that” issues when auditing a website.

Uppercase redirect rule

Any Tech SEO worth their salt will know about the obvious redirect rules and where they should be applied e.g. http to https pages, trailing slash to non-trailing slash etc.

However, when was the last time you checked that all page URLs force the lowercase?

As URL is the unique identifier for a webpage, therefore www.example.com/category and www.example.com/Category are considered different pages, even though the content is the same.

This can create duplicate content issues within search engines, so lowercase URLs should be forced.

Pagination problems

If the website you manage has category pagination, then the URL pattern is often as follows:


However, the ?p=1 page, is a duplicate of the canonical www.example.com/mattresses/memory-foam-mattresses page, in this example.

Therefore, this ?p=1 duplicate should be 301 redirected to the www.example.com/mattresses/memory-foam-mattresses page, with all internal links updated to reflect the change.

Another easy-to-forget pagination problem is with canonical tags. Often the ?p=2, ?p=3 pages are referencing the first paginated page in their canonical tags.

Referencing the first page on all paginated pages tells search engines the pages are the same as the first page, meaning crawlers may not crawl the other pages or follow the links to the products from them. Make them self-referencing!

Use Meta noindex Tag Instead of robots.txt

I see the “No information is available for this page” message ALL THE TIME when looking at indexed pages within Google:

Use Meta noindex Tag Instead of robots.txt

This is because the webmaster has tried to remove a page from the index by adding a disallow directive against it in the robots.txt.

However, these pages can still be found via searches, so it is better to remove them from the robots.txt and add a noindex tag between the HEAD tags of the pages that shouldn’t be indexed.

Remove Meta nofollow from Select Pages

Now, this obviously doesn’t apply to all pages that have a nofollow tag, but many nofollow tags are often added with a noindex tag by mistake. Removing the nofollow instruction will improve the flow of link authority and crawler access around the site.

Although search engines might not need to index these pages, pages they link to could be indexed, therefore the links should be followed.

Dev Site in the Index

I love finding a staging site still indexed when auditing a potential client's website. It shows a lack of developer care and can be a great way to get the client on your side early in a relationship.

The best way to block a test environment from getting indexed is to password-protect or restrict based on IP range.

Some quick methods to check if your dev site is being indexed within Google is with advanced search operators such as:

- Site:dev.example.com
- Site:staging.example.com
- Site:prod.example.com
- Site:example.com inurl:test
- Site:example.com -inurl:www.

Review your XML Sitemap

XML sitemaps often get forgotten about if they aren’t automatically generated. Here are 3 quick checks that take less than 5 minutes combined:

Check 1 – Is the XML Sitemap being linked to in the robots.txt?

Search engine crawlers will always look for a robots.txt file in the root folder of a website. Therefore, adding a link to your XML sitemap within the robots.txt points all major search engines to its location.

Check 2 – Has the XML Sitemap been added to Google Search Console?

Submitting XML sitemaps to Google Search Console ensures Google can find and crawl them. Once submitted, you can see if Google is successfully able to discover your page URLs:

Has the XML Sitemap been added to Google Search Console?

Check 3 – Are there any errors in the XML sitemap?

URLs in sitemaps should not redirect or return errors – every page should return a 200 OK status code. Search engines may ignore sitemaps completely if they contain too many errors.

You can discover if your sitemap has any errors by looking in Google Search Console once the XML sitemap has been submitted. Alternatively, you can crawl an XML sitemap in Screaming Frog by selecting Mode > List > Upload > Download XML Sitemap.

Remove Old Content from Index

When was the last time you looked at your dated content, such as competition pages for contests that ended in 2017?

Reviewing which content is being indexed within Google can lead to technical SEO benefits, like improved crawling and indexation improvements, so it is an important area to consider.

For instance, if there are hundreds of out-of-date blog posts being indexed then search engines could be wasting their crawl budget visiting these pages, when they could be processing more important pages.

Remove Old Content from Index

I like to find these dated blog posts with advanced search operators. Some of my favourites include:

- Site:example.com inurl:Christmas
- Site:example.com inurl:competition
- Site:example.com inurl:award
- Site:example.com inurl:august
- Site:example.com inurl:september
- Site:example.com inurl:2016
- Site:example.com inurl:2017
- Etc.

Try Ranktracker for FREE