Technical SEO Audit: What Google Sees (And What It Doesn't)

Send me an e-mail

Most websites leak rankings through technical errors no one has bothered to find. In one audit I conducted, nearly half of all pages were set to noindex. No one noticed for months while organic traffic steadily bled out. A single misconfigured robots.txt or 49% of pages set to noindex can silently kill months of content work. This guide shows you exactly how to run a technical SEO audit that surfaces what Google cannot see.

What is a Technical SEO Audit (And why most teams skip it)

Technical SEO is the foundation of the search hierarchy. If you look at the Moz SEO pyramid, technical health sits at the base. Without crawlability and indexation, content and links amplify nothing because bots cannot reach the site. In my 5+ years of navigating SEO shifts at agencies like Groove Digital, I have seen that teams often deprioritise technical health because results are less visible than a new blog post or a high-tier backlink.

Read more about: securing stakeholder buy-in.

However, a website audit is not a luxury. It is a maintenance layer that keeps your strategy alive. Most teams focus on content because it feels more creative. Technical SEO feels like "fixing the plumbing." Yet, a quarterly audit cadence is the bare minimum for any professional site. For e-commerce platforms with thousands of SKUs, a monthly crawl is essential to catch expiring products or parameter URL bloat. If the foundation is cracked, your skyscraper will never reach page one.

What can you audit on a website? The Full Checklist

Before diving into the data, you need a mental model of what you are actually checking. A comprehensive SEO audit covers four main dimensions: performance, technical SEO, content, and backlinks. Using the right SEO audit tools for each task is critical for efficiency.

  • Performance: Use PageSpeed Insights for load time and GA4 to monitor fluctuations in organic traffic.
  • Technical SEO: Screaming Frog is the industry standard for auditing redirects, crawl depth, and indexation status.
  • Content: You can audit titles, meta descriptions, H1/H2 tags, and image alt text directly within a Screaming Frog crawl.
  • Backlinks: Use Ahrefs for broken link detection and to ensure your external authority is not pointing to 404 pages.
Tool Primary Task
Screaming Frog Technical health, crawlability, and on-page metadata.
Ahrefs Backlink profile and broken external link detection.
Google Search Console Indexation confirmation and keyword-level performance.
PageSpeed Insights Core Web Vitals and loading performance.

Learn more about: Google Tag Manager for SEO.

Crawling vs. Indexing: What is actually the difference?

Crawling and indexing are distinct processes. Crawling is the act of Googlebot following links to discover your pages. Indexing is the act of storing that content in Google’s database to be ranked. A page can be crawled but not indexed. It can also be blocked from both.

The robots.txt file controls crawling at the server level. The meta robots tag controls indexing at the page level. If you want a page to be "invisible" but still pass link equity, you use "noindex, follow." This tells Google not to show the page in results but to keep following its internal links to other parts of your site.

How robots.txt works in practice

To check for blocked pages, open Screaming Frog and head to the Response Codes tab. Filter by Blocked by robots.txt. You might find that /wp-admin/ is blocked, which is legitimate. However, if your primary service pages appear here, Googlebot is being turned away at the door.

Meta Robots: noindex, follow Explained

In the HTML, this looks like <meta name="robots" content="noindex, follow">. Use this for thank-you pages or internal search result pages. It prevents "thin content" from entering the index while allowing bots to navigate your site structure.

Multilingual and multi-regional sites add another layer: hreflang. Get it wrong and Googlebot serves the Dutch page to a German user. If you're auditing across borders, my hreflang implementation guide is the next logical step.

How to find and fix pages blocked by robots.txt

A robots.txt audit prevents wasted crawl budget. When Googlebot spends time trying to access blocked URLs, it has less time for your "money pages." In Screaming Frog, go to the Response Codes tab and filter for Blocked by robots.txt.

Check the Matched Robots.txt Line column. This tells you exactly which rule is causing the block. Triage these into two categories: intentional and accidental. If you see blog posts or product categories here, you are killing your own rankings. Fix these by updating your robots.txt file on the server to allow access to those specific paths.

Noindex Audit: How to find pages Google is ignoring

Finding noindex pages is done via the Directives tab in Screaming Frog. Filter for Noindex to see what is hidden from the search results. On a live site I audited, 73 out of 148 pages were noindexed. That is 49% of the site invisible to Google.

Common culprits include WordPress tag archives or staging environments that were accidentally left in "noindex" mode after a launch. If you find indexable content here, remove the meta robots tag immediately. Every day a page is set to noindex is a day it generates zero ROI.

Duplicate content: The silent rankings killer

Duplicate content SEO issues occur when search engines find the same content on multiple URLs. This creates confusion. Google does not know which version to rank, so it often ranks none of them well. This also wastes your crawl budget.

Common variants include HTTPS vs. HTTP or www vs. non-www. In e-commerce, product pages with minimal variations in size or colour are the most frequent offenders. Google splits link equity across these duplicates instead of consolidating it into one strong URL.

How to run a Duplicate Content check in Screaming Frog

Navigate to Configuration > Content > Duplicates. Enable Near Duplicates and run a Crawl Analysis. Once finished, check the Content tab. If you see a 94% similarity match, you have a near-duplicate issue that needs addressing.

How to fix Duplicate Content

Use canonical tags for near-duplicates to tell Google which URL is the "master" version. Use 301 redirects for structural duplicates like HTTP to HTTPS. For clusters of thin, similar pages, consider content consolidation into one comprehensive guide.

Content audit in Screaming Frog: Titles, Descriptions, and Alt Text

A content audit ensures your on-page elements work for both crawlers and users. Screaming Frog surfaces titles, meta descriptions, H1 tags, image alt text, and broken links in a single crawl — no manual page-by-page checking required.

Three issues consistently damage performance on otherwise well-built sites:

  • Missing alt text removes a ranking signal Google uses to interpret image content critical for product-heavy or visual-first pages
  • Broken internal links create dead ends for crawlers, waste crawl budget, and signal poor site maintenance
  • Even non-indexable pages benefit from optimised titles users still land on them, and first impressions matter

Page Title optimisation: Length, Pixel Width, and Click Intent

In 2026, pixel width is more accurate than character count. Character widths vary, but the SERP space is fixed. Aim for 30–60 characters or 200–554 pixels for your SEO title length.

Lead with your primary keyword. Close with your brand using a pipe or a colon. For example: "SEO Content Guidelines: All You Need to Know in 2026 | Sahardid". If your title does not match the page content, Google might rewrite it. This is a warning sign that you are over-optimising or missing the user's search intent.

Meta Description optimisation: Writing for clicks

Meta descriptions are not a direct ranking factor. They are click-through rate (CTR) levers. Use 70–155 characters or 400–1005 pixels.

Address active intent first by using the primary keyword. Address passive intent next by mentioning tools, tips, or examples. Use Google Search Console (GSC) to monitor performance. If a page has high impressions but a CTR below 3%, your description likely needs a rewrite to better match the search intent.

Most SEOs stop here. The ones who don't add schema markup and rich snippets and that's where the real CTR gap opens up.

How to measure performance in Google Search Console

To validate your SERP copy, use GSC. Filtering by page level is often too noisy because it includes every random query. Instead, filter by Page + Main Keyword.

This isolates how well you match the primary intent. A high-impression, low-CTR keyword is your priority target for a title or description rewrite. For branded navigational queries, you should aim for a benchmark of 45% or higher CTR when sitting in positions one or two.

Organising your Technical SEO Audit in Google Sheets

Raw data is useless without organisation. Export the SERP Summary from Screaming Frog and import it into Google Sheets. Use conditional formatting to flag issues.

  • Title pixel width > 554px: Set to red.
  • Title pixel width < 200px: Set to orange.

Add a GSC CTR column next to each URL for context. Prioritise your fixes. Broken links and noindex errors must be addressed first. Title and description rewrites come second. This turns a crawl into a client-ready SEO audit report.

Dev teams ignore spreadsheets. They act on priorities. Once your data is in Google Sheets, turn your findings into a structured SEO roadmap what gets fixed this sprint, what waits for Q4, and why.

Conclusion

Technical SEO is not a one-time task. It is the maintenance layer that keeps everything else working. As Google's crawl behaviour evolves in 2026 with AI Overviews and reduced crawl budgets, a clean technical foundation is the non-negotiable baseline. Audit quarterly. Prioritise crawlability. Let your content and links do the rest.

Frequently Asked Questions

What is a technical SEO audit?

A systematic review of a website's crawlability, indexation, and on-page technical elements to identify issues blocking visibility.

How often should I run a technical SEO audit?

A minimum of every quarter for most sites. Monthly audits are recommended for e-commerce sites with large page counts.

What does Screaming Frog check in a website audit?

It checks for broken links, redirect chains, noindex pages, duplicate content, and missing metadata.

What's the difference between crawling and indexing?

Crawling is discovery; indexing is storage and ranking. A page can be crawled but excluded from the index.

How do I fix duplicate content issues?

Use canonical tags, 301 redirects, or content consolidation.

What is a good meta description length in 2026?

Between 70–155 characters or 400–1005 pixels for visual accuracy

About the Author

Mohamed Sahardid Senior SEO & Growth Strategist

Mohamed Sahardid is a senior SEO strategist based in Amsterdam, with agency experience at Groove Digital and Chase Marketing. He specialises in technical SEO, topical authority, and Answer Engine Optimisation helping brands get found in both traditional search and AI-generated results.

mo@sahardid.com · sahardid.com