How to Perform a Technical SEO Audit: A 10-Step Guide (2024)


A technical SEO audit analyzes the technical aspects of a website related to search engine optimization. It ensures search engines like Google can crawl, index, and rank pages on your site.

In a technical SEO audit, you’ll look at (and fix) things that could:

  • Slow down your site
  • Make it difficult for search engines to understand your content
  • Make it hard for your pages to appear in search results
  • Affect how users interact with your site on different devices
  • Impact your site’s security
  • Create duplicate content issues
  • Cause navigation problems for users and search engines
  • Prevent important pages from being found

Identifying and fixing such technical issues help search engines better understand and rank your content. Which can mean improved organic search visibility and traffic over time.

How to Perform a Technical SEO Audit

You’ll need two main tools for a technical site audit:

  1. Google Search Console
  2. A crawl-based tool, like Semrush’s Site Audit

If you haven’t used Search Console before, check out our beginner’s guide. We’ll discuss the tool’s various reports below.

And if you’re new to Site Audit, sign up for free account to follow along with this guide. 

The Site Audit tool scans your website and provides data about each page it crawls. The report it generates shows you a variety of technical SEO issues.

In a dashboard like this: 

Site Audit overview showing site health, errors, warnings and notices, a breakdown of crawled pages, and thematic reports

To set up your first crawl, create a project.

"Create project" window on Site Audit with a domain and project name entered

Next, head to the Site Audit tool and select your domain.

"Projects" page on "Site Audit" with a domain highlighted and clicked

The “Site Audit Settings” window will pop up. Here, configure the basics of your first crawl. Follow this detailed setup guide for help.

Site Audit settings page to set crawl scope, source, and limit of checked pages

Finally, click “Start Site Audit.” 

Site Audit settings page with the "Start Site Audit" button clicked

After the tool crawls your site, it generates an overview of your site’s health. 

"Site Health" score on Site Audit overview highlighted

This metric grades your website health on a scale from 0 to 100. And how you compare with other sites in your industry.

Your site issues are ordered by severity through the “Errors,” “Warnings,” and “Notices” categories. Or focus on specific areas of technical SEO with “Thematic Reports.”

"Thematic Reports" and "Errors, Warnings, and Notices" on "Site Audit" overview highlighted

Toggle to the Issues tab to see a complete list of all site issues. Along with the number of affected pages.

"Issues" tab on Site Audit showing a list of warnings like too much text within the title tags, don't have meta descriptions, have a low word count, etc.

Each issue includes a “Why and how to fix it” link.

“Why and how to fix it” clicked showing a short description of an issue, tips on how to fix it, and useful links to relevant tools

The issues you find here will fit into one of two categories, depending on your skill level:

  • Issues you can fix on your own
  • Issues a developer or system administrator might need to help you fix

Conduct a technical SEO audit on any new site you work with. Then, audit your site at least once per quarter (ideally monthly). Or whenever you see a decline in rankings.

1. Spot and Fix Crawlability and Indexability Issues

Crawlability and indexability are a crucial aspect of SEO. Because Google and other search engines must be able to crawl and index your webpages in order to rank them.

Google’s bots crawl your site by following links to find pages. They read your content and code to understand each page. 

Google then stores this information in its index—a massive database of web content. 

When someone performs a Google search, Google checks its index to return relevant results. 

how search engines work: from publishing content, spiders crawling the site, Google indexing the page, to showing up on the SERP

To check if your site has any crawlability or indexability issues, go to the Issues tab in Site Audit. 

Then, click “Category” and select “Crawlability.” 

Site Audit issues with the “Category” drop-down opened and and “Crawlability” clicked

Repeat this process with the “Indexability” category.

Issues connected to crawlability and indexability will often be at the top of the results in the “Errors” section. Because they’re often more serious. We’ll cover several of these issues. 

"Errors" on Site Audit showing the most serious website issues like broken internal links, 5xx status code errors, 4xx status code errors, etc.

Now, let’s look at two important website files—robots.txt and sitemap.xml—that have a huge impact on how search engines discover your site.

Spot and Fix Robots.txt Issues

Robots.txt is a website text file that tells search engines which pages they should or shouldn’t crawl. It can usually be found in the root folder of the site: https://domain.com/robots.txt. 

A robots.txt file helps you:

  • Point search engine bots away from private folders
  • Keep bots from overwhelming server resources
  • Specify the location of your sitemap

A single line of code in robots.txt can prevent search engines from crawling your entire site. Make sure your robots.txt file doesn’t disallow any folder or page you want to appear in search results.

To check your robots.txt file, open Site Audit and scroll down to the “Robots.txt Updates” box at the bottom.

"Robots.txt Updates" box highlighted on Site Audit overview

Here, you’ll see if the crawler has detected the robots.txt file on your website.

If the file status is “Available,” review your robots.txt file by clicking the link icon next to it. 

Or, focus only on the robots.txt file changes since the last crawl by clicking the “View changes” button. 

"Robots.txt Updates" box highlighted with the “View changes” button clicked

Further reading: Reviewing and fixing the robots.txt file requires technical knowledge. Always follow Google’s robots.txt guidelines. Read our guide to robots.txt to learn about its syntax and best practices. 

To find further issues, open the “Issues” tab and search “robots.txt.” 

"Issues" tab on Site Audit clicked and “robots.txt” entered in the search bar

Some issues include:

  • Robots.txt file has format errors: Your robots.txt file might have mistakes in its setup. This could accidentally block important pages from search engines or allow access to private content you don’t want shown.
  • Sitemap.xml not indicated in robots.txt: Your robots.txt file doesn’t mention where to find your sitemap. Adding this information helps search engines find and understand your site structure more easily.
  • Blocked internal resources in robots.txt: You might be blocking important files (like CSS or JavaScript) that search engines need to properly view and understand your pages. This can hurt your search rankings.
  • Blocked external resources in robots.txt: Resources from other websites that your site uses (like CSS, JavaScript, and image files) might be blocked. This can prevent search engines from fully understanding your content.

Click the link highlighting the found issues. 

a list of issues showing on Site Audit for the “robots.txt” search with "Robots.txt file has format errors" clicked

Inspect them in detail to learn how to fix them.

"Robots.txt Updates" showing file status, changes, and columns for user-agent, event, and rule

Further reading: Besides the robot.txt file, there are two other ways to provide instructions for search engine crawlers: the robots meta tag and x-robots tag. Site Audit will alert you of issues related to these tags. Learn how to use them in our guide to robots meta tags.

Spot and Fix XML Sitemap Issues

An XML sitemap is a file that lists all the pages you want search engines to index and rank.

Review your XML sitemap during every technical SEO audit to ensure it includes all pages you want to rank.

Also check that the sitemap doesn’t include pages you don’t want in the SERPs. Like login pages, customer account pages, or gated content.

Next, check whether your sitemap works correctly.

The Site Audit tool can detect common sitemap-related issues, such as:

  • Format errors: Your sitemap has mistakes in its setup. This could confuse search engines, causing them to ignore your sitemap entirely.
  • Incorrect pages found: You’ve included pages in your sitemap that shouldn’t be there, like duplicate content or error pages. This can waste your crawl budget and confuse search engines.
  • File is too large: Your sitemap is bigger than search engines prefer. This might lead to incomplete crawling of your site.
  • HTTP URLs in sitemap.xml for HTTPS site: Your sitemap lists unsecure versions of your pages on a secure site. This mismatch could mislead search engines.
  • Orphaned pages: You’ve included pages in your sitemap that aren’t linked from anywhere else on your site. This could waste the crawl budget on potentially outdated or unimportant pages.

To find and fix these issues, go to the “Issues” tab and type “sitemap” in the search field:

“Issues” tab on Site Audit with "sitemap" entered in the search bar

You can also use Google Search Console to identify sitemap issues.

Visit the “Sitemaps” report to submit your sitemap to Google, view your submission history, and review any errors. 

Find it by clicking “Sitemaps” under the “Indexing” section.

Google Search Console menu with “Sitemaps” under the “Indexing” section clicked

If you see “Success” listed next to your sitemap, there are no errors. But the other two statuses—“Has errors” and “Couldn’t fetch”—indicate a problem.

"Submitted sitemaps" on GSC with columns for sitemap, type, **** submitted, **** last read, status, and discovered URLs

If there are issues, the report will flag them individually. Follow Google’s troubleshooting guide to fix them. 

Further reading: If your site doesn’t have a sitemap.xml file, read our guide on how to create an XML sitemap

2. Audit Your Site Architecture

Site architecture refers to the hierarchy of your webpages and how they are connected through links. Organize your website so it’s logical for users and easy to maintain as your website grows.

Good site architecture is important for two reasons:

  1. It helps search engines crawl and understand the relationships between your pages
  2. It helps users navigate your site

Let’s consider three key aspects of site architecture. And how to analyze them with the technical SEO audit tool.

Site Hierarchy

Site hierarchy (or site structure) is how your pages are organized into subfolders.

To understand site’s hierarchy, navigate to the “Crawled Pages” tab in Site Audit.

navigating to the "Crawled Pages" tab on Site Audit

Then, switch the view to “Site Structure.”

"Site Structure" view on "Crawled Pages" showing an overview of your website’s subdomains and subfolders

You’ll see your website’s subdomains and subfolders. Review them to make sure the hierarchy is organized and logical.

Aim for a flat site architecture, which looks like this:

a flat site architecture where users can access pages from your homepage within three clicks

Ideally, it should only take a user three clicks to find the page they want from your homepage.

When it takes more than three clicks to navigate your site, its hierarchy is too deep. Search engines consider pages deep in the hierarchy to be less important or relevant to a search query.

To ensure all your pages fulfill this requirement, stay within the “Crawled Pages” tab and switch back to the “Pages” view.

"Pages" view on “Crawled Pages” with columns for page URL, unique pageviews, crawl depth, issues, etc.

Then, click “More filters” and select the following parameters: “Crawl Depth” is “4+ clicks.”

“More filters” on "Crawled Pages" clicked and “Crawl Depth” is “4+ clicks” set as the parameter

To fix this issue, add internal links to pages that are too deep in the site’s structure. 

Your site’s navigation (like menus, footer links, and breadcrumbs) should make it easier for users to navigate your site. 

This is an important pillar of good website architecture.

Your navigation should be:

  • Simple. Try to avoid mega menus or non-standard names for menu items (like “Idea Lab” instead of “Blog”)
  • Logical. It should reflect the hierarchy of your pages. A great way to achieve this is to use breadcrumbs.

Breadcrumbs are a secondary navigation that shows users their current location on your site. Often appearing as a row of links at the top of a page. Like this:

breadcrumb navigation example from the men’s jeans page on Nordstrom

Breadcrumbs help users understand your site structure and easily move between levels. Improving both user experience and SEO.

No tool can help you create user-friendly menus. You need to review your website manually and follow UX best practices for navigation

URL Structure

Like a website’s hierarchy, a site’s URL structure should be consistent and easy to follow. 

Let’s say a website visitor follows the menu navigation for *****’ shoes:

Homepage > Children > ***** > Footwear

The URL should mirror the architecture: domain.com/children/*****/footwear

Some sites should also consider using a URL structure that shows a page or website is relevant to a specific country. For example, a website for Canadian users of a product may use either “domain.com/ca” or “domain.ca.”

Lastly, make sure your URL slugs are user-friendly and follow best practices. 

Site Audit identifies common issues with URLs, such as:

  • Use of underscores in URLs: Using underscores (_) instead of hyphens (-) in your URLs can confuse search engines. They might see words connected by underscores as a single word, potentially affecting your rankings. For example, “blue_shoes” could be read as “blueshoes” instead of “blue shoes”.
  • Too many parameters in URLs: Parameters are URL elements that come after a question mark, like “?color=blue&size=large”. They help with tracking. Having too many can make your URLs long and confusing, both for users and search engines.
  • URLs that are too long: Some browsers might have trouble processing URLs that exceed 2,000 characters. Short URLs are also easier for users to remember and share.
"Warnings" on Site Audit like too many parameters, have underscores in the URL, etc. highlighted

3. Fix Internal Linking Issues

Internal links point from one page to another within your domain.

Internal links are an essential part of a good website architecture. They distribute link equity (also known as “link juice” or “authority”) across your site. Which helps search engines identify important pages.

As you improve your site’s structure, check the health and status of its internal links.

Refer back to the Site Audit report and click “View details” under your “Internal Linking” score. 

"Internal linking" on Site Audit highlighted and clicked

In this report, you’ll see a breakdown of your site’s internal link issues.

"Internal Linking" report on Site Audit showing a breakdown of a site's internal link issues

Broken internal links—links that point to pages that no longer exist—are a common internal linking mistake. And are fairly easy to fix. 

Click the number of issues in the “Broken internal links” error on your “Internal Link Issues” report. And manually update the broken links in the list. 

Broken internal links error on the "Internal Linking" report highlighted and clicked

Another easy fix is orphaned pages. These are pages with no links pointing to them. Which means you can’t gain access to them via any other page on the same website.

Check the “Internal Links” bar graph to look for pages with zero links. 

Internal Links bar graph showing a page with zero links highlighted

Add at least one internal link to each of these pages. 

Use the “Internal Link Distribution” graph to see the distribution of your pages according to their Internal LinkRank (ILR).

ILR shows how strong a page is in terms of internal linking. The closer to 100, the stronger a page.

Internal link distribution report showing a breakdown of a site's pages based on their internal link strength

Use this metric to learn which pages could benefit from additional internal links. And which pages you can use to distribute more link equity across your domain. 

But don’t continue fixing issues that could have been avoided. Follow these internal linking best practices to avoid issues in the future:

  • Make internal linking part of your content creation strategy
  • Every time you create a new page, link to it from existing pages
  • Don’t link to URLs that have redirects (link to the redirect destination instead)
  • Link to relevant pages and use relevant anchor text
  • Use internal links to show search engines which pages are important
  • Don’t use too many internal links (use common sense here—a standard blog post likely doesn’t need 50 internal links)
  • Learn about nofollow attributes and use them correctly

4. Spot and Fix Duplicate Content Issues

Duplicate content means multiple webpages contain identical or nearly identical content. 

It can lead to several problems, including:

  • SERPs displaying an incorrect version of your page
  • The most relevant pages not performing well in SERPs
  • Indexing problems on your site
  • Splitting your page authority between duplicate versions
  • Increased difficulty in tracking your content’s performance

Site Audit flags pages as duplicate content if their content is at least 85% identical. 

"duplicate content issues" on Site Audit errors highlighted and clicked

Duplicate content can happen for two common reasons:

  1. There are multiple versions of URLs
  2. There are pages with different URL parameters

Multiple Versions of URLs

For example, a site may have:

  • An HTTP version
  • An HTTPS version
  • A www version
  • A non-www version

For Google, these are different versions of the site. So if your page runs on more than one of these URLs, Google considers it a duplicate.

To fix this issue, select a preferred version of your site and set up a sitewide 301 redirect. This will ensure only one version of each page is accessible.

URL Parameters

URL parameters are extra elements of a URL used to filter or sort website content. They’re commonly used for product pages with slight changes (e.g., different color variations of the same product).

You can identify them because by the question mark and equal sign.

the URL parameter on a product page URL of "Mejuri" highlighted

Because URLs with parameters have almost the same content as their counterparts without parameters, they can often be identified as duplicates. 

Google usually groups these pages and tries to select the best one to use in search results. Google will typically identify the most relevant version of the page and display that in search results—while consolidating ranking signals from the duplicate versions.

Nevertheless, Google recommends these actions to reduce potential problems:

  • Reduce unnecessary parameters
  • Use canonical tags pointing to the URLs with no parameters

Avoid crawling pages with URL parameters when setting up your SEO audit. To ensure the Site Audit tool only crawls pages you want to analyze—not their versions with parameters.

Customize the “Remove URL parameters” section by listing all the parameters you want to ignore:

"Remove URL parameters" on Site Audit Settings with a list of parameters entered in the input box

To access these settings later, click the settings (gear) icon in the top-right corner, then click “Crawl sources: Website” under the Site Audit settings. 

the gear icon on Site Audit clicked and "Crawl sources: Website" selected from the drop-down

5. Audit Your Site Performance

Site speed is a crucial aspect of the overall page experience and has long been a Google ranking factor.

When you audit a site for speed, consider two data points:

  1. Page speed: How long it takes one webpage to load
  2. Site speed: The average page speed for a sample set of page views on a site

Improve page speed, and your site speed improves.

This is such an important task that Google has a tool specifically made to address it: PageSpeed Insights

"Core Web Vitals Assessment" on "PageSpeed Insights" showing metrics like LCP, INP, CLS, FCP, FID, and TTFB

A handful of metrics influence PageSpeed scores. The three most important ones are called Core Web Vitals

They include:

  • Largest Contentful Paint (LCP): measures how fast the main content of your page loads 
  • Interaction to Next Paint (INP): measures how quickly your page responds to user interactions
  • Cumulative Layout Shift (CLS): measures how visually stable your page is 
a breakdown of LCP, INP, and CLS into three categories: good, needs improvement, and poor

PageSpeed Insights provides details and opportunities to improve your page in four main areas:

  • Performance
  • Accessibility
  • Best Practices
  • SEO
PageSpeed Insights with scores for a site's performance, accessibility, best practices, and SEO

But PageSpeed Insights can only analyze one URL at a time. To get the sitewide view, use Semrush’s Site Audit.

Head to the “Issues” tab and select the “Site Performance” category.

Here, you can see all the pages a specific issue affects—like slow load speed. 

"Site Performance" selected as the category on Site Audit Issues

There are also two detailed reports dedicated to performance—the “Site Performance” report and the “Core Web Vitals” report. 

Access both from the Site Audit Overview.

thematic reports on Site Audit with “Site Performance” and “Core Web Vitals” highlighted

The “Site Performance” report provides an additional “Site Performance Score.” Or a breakdown of your pages by their load speed and other useful insights.

Site Performance report showing a breakdown of a site's pages by load speed on the left and performance issues on the right

The Core Web Vitals report will break down your Core Web Vitals metrics based on 10 URLs. Track your performance over time with the “Historical Data” graph.

Or edit your list of analyzed pages so the report covers various types of pages on your site (e.g., a blog post, a landing page, and a product page).

Click “Edit list” in the “Analyzed Pages” section.

"Edit list" on top of the “Analyzed Pages” section clicked

Further reading: Site performance is a broad topic and one of the most important aspects of technical SEO. To learn more about the topic, check out our page speed guide, as well as our detailed guide to Core Web Vitals

6. Discover Mobile-Friendliness Issues

As of January 2024, more than half (60.08%) of web traffic happens on mobile devices.

And Google primarily indexes the mobile version of all websites over the desktop version. (Known as mobile-first indexing.) 

So ensure your website works perfectly on mobile devices. 

Use Google’s Mobile-Friendly Test to quickly check mobile usability for specific URLs.

And use Semrush to check two important aspects of mobile SEO: viewport meta tag and AMPs. 

Just select the “Mobile SEO” category in the “Issues” tab of the Site Audit tool. 

"Mobile SEO" selected as the category on Site Audit Issues showing a list of related issues

A viewport meta tag is an HTML tag that helps you scale your page to different screen sizes. It automatically alters the page size based on the user’s device when you have a responsive design.

Another way to improve the site performance on mobile devices is to use Accelerated Mobile Pages (AMPs), which are stripped-down versions of your pages.

AMPs load quickly on mobile devices because Google runs them from its cache rather than sending requests to your server.

If you use AMPs, audit them regularly to make sure you’ve implemented them correctly to boost your mobile visibility.

Site Audit will test your AMPs for various issues divided into three categories:

  1. AMP HTML issues
  2. AMP style and layout issues
  3. AMP templating issues

7. Spot and Fix Code Issues

Regardless of what a webpage looks like to human eyes, search engines only see it as a bunch of code.

So, it’s important to use proper syntax. And relevant tags and attributes that help search engines understand your site.

During your technical SEO audit, monitor different parts of your website code and markup. Including HTML (which includes various tags and attributes), JavaScript, and structured data. 

Let’s dig into these. 

Meta Tag Issues

Meta tags are text snippets that provide search engine bots with additional data about a page’s content. These tags are present in your page’s header as a piece of HTML code.

We’ve already covered the robots meta tag (related to crawlability and indexability) and the viewport meta tag (related to mobile-friendliness). 

You should understand two other types of meta tags:

  1. Title tag: Indicates the title of a page. Search engines use title tags to form the clickable blue link in the search results. Read our guide to title tags to learn more.
  2. Meta description: A brief description of a page. Search engines use it to form the snippet of a page in the search results. Although not directly tied to Google’s ranking algorithm, a well-optimized meta description has other potential SEO benefits like improving click-through rates and making your search result stand out from competitors.
title tag and meta description for a SERP listing on Google highlighted

To see issues related to meta tags in your Site Audit report, select the “Meta tags” category in the “Issues” tab.

"Meta tags" selected as the category on Site Audit Issues showing a list of related issues

Here are some common meta tag issues you might find:

  • Missing title tags: A page without a title tag may be seen as low quality by search engines. You’re also missing an opportunity to tell users and search engines what your page is about.
  • Duplicate title tags: When multiple pages have the same title, it’s hard for search engines to determine which page is most relevant for a search query. This can hurt your rankings.
  • Title tags that are too long: If your title exceeds 70 characters, it might get cut off in search results. This looks unappealing and might not convey your full message.
  • Title tags that are too short: Titles with 10 characters or less don’t provide enough information about your page. This limits your ability to rank for different keywords.
  • Missing meta descriptions: Without a meta description, search engines might use random text from your page as the snippet in search results. This could be unappealing to users and reduce click-through rates.
  • Duplicate meta descriptions: When multiple pages have the same meta description, you’re missing chances to use relevant keywords and differentiate your pages. This can confuse both search engines and users.
  • Pages with a meta refresh tag: This outdated technique can cause SEO and usability issues. Use proper redirects instead.

Canonical Tag Issues

Canonical tags are used to point out the “canonical” (or “main”) copy of a page. They tell search engines which page needs to be indexed in case there are multiple pages with duplicate or similar content. 

A canonical URL tag is placed in the <head> section of a page’s code and points to the “canonical” version.

It looks like this:

<link rel="canonical" href="https://www.domain.com/the-canonical-version-of-a-page/" />

A common canonicalization issue is that a page has either no canonical tag or multiple canonical tags. Or, of course, a broken canonical tag. 

The Site Audit tool can detect all of these issues. To only see the canonicalization issues, go to “Issues” and select the “Canonicalization” category in the top filter.

"Canonicalization" selected as the category on Site Audit Issues showing a list of related issues

Common canonical tag issues include:

  • AMPs with no canonical tag: If you have both AMP and non-AMP versions of a page, missing canonical tags can lead to duplicate content issues. This confuses search engines about which version to show in the results.
  • No redirect or canonical to HTTPS homepage from HTTP version: When you have both HTTP and HTTPS versions of your homepage without proper direction, search engines struggle to know which one to prioritize. This can split your SEO efforts and hurt your rankings.
  • Pages with a broken canonical link: If your canonical tag points to a non-existent page, you’re wasting the crawl budget and confusing search engines.
  • Pages with multiple canonical URLs: Having more than one canonical tag on a page gives conflicting directions. Search engines might ignore all of them or pick the wrong one, potentially hurting your SEO results.

Hreflang Attribute Issues

The hreflang attribute denotes the target region and language of a page. It helps search engines serve the correct variation of a page based on the user’s location and language preferences.

If your site needs to reach audiences in more than one country, use hreflang attributes in <link> tags.

Like this:

hreflang attributes being used in <link> tag shown on a site's source code

To audit your hreflang annotations, go to the “International SEO” thematic report in Site Audit. 

"International SEO" under "Thematic Reports" on Site Audit clicked

You’ll see a comprehensive overview of the hreflang issues on your site:

"348 issues" next to "Hreflang conflicts within page source code" on the International SEO report clicked

And a detailed list of pages with missing hreflang attributes on the total number of language versions your site has.

a list of pages with missing hreflang attributes on the total number of language versions a site has on Site Audit

Common hreflang issues include:

  • Pages with no hreflang and lang attributes: Without these, search engines can’t determine the language of your content or which version to show users.
  • Hreflang conflicts within page source code: Contradictory hreflang information confuses search engines. This can lead to the wrong language version appearing in search results.
  • Issues with hreflang values: Incorrect country or language codes in your hreflang attributes prevent search engines from properly identifying the target audience for your content. This can lead to your pages being shown to the wrong users.
  • Incorrect hreflang links: Broken or redirecting hreflang links make it difficult for search engines to understand your site’s language structure. This can result in inefficient crawling and improper indexing of your multilingual content.
  • Pages with hreflang language mismatch: When your hreflang tag doesn’t match the actual language of the page, it’s like false advertising. Users might land on pages they can’t understand.

Fixing these issues helps ensure that your international audience sees the right content in search results. Which improves user experience and potentially boosts your global SEO ROI.

JavaScript Issues

JavaScript is a programming language used to create interactive elements on a page. 

Search engines like Google use JavaScript files to render the page. If Google can’t get the files to render, it won’t index the page properly.

The Site Audit tool detects broken JavaScript files and flags the affected pages.

a list of issues showing for the term "javascript" like slow load speed, broken JavaScript and CSS files, etc.

It can also show other JavaScript-related issues on your website. Including:

  • Unminified JavaScript and CSS files: These files contain unnecessary code like comments and extra spaces. Minification removes this excess, reducing file size without changing functionality. Smaller files load faster.
  • Uncompressed JavaScript and CSS files: Even after minification, these files can be compressed further. Compression reduces file size, making them quicker to download.
  • Large total size of JavaScript and CSS: If your combined JS and CSS files exceed 2 MB after minification and compression, they can still slow down your page. This large size leads to poor UX and potentially lower search rankings.
  • Uncached JavaScript and CSS files: Without caching, browsers must download these files every time a user visits your site. This increases load time and data usage for your visitors.
  • Too many JavaScript and CSS files: Using more than 100 files increases the number of server requests, slowing down your page load time
  • Broken external JavaScript and CSS files: When files hosted on other sites don’t work, it can cause errors on your pages. This affects both user experience and search engine indexing.

Addressing these issues can improve your site’s performance, user experience, and search engine visibility.

To check how Google renders a page that uses JavaScript, go to Google Search Console and use the “URL Inspection Tool.”

Enter your URL into the top search bar and hit enter.

a URL entered on the "URL Inspection" tool on Google Search Console

Then, test the live version of the page by clicking “Test Live URL” in the top-right corner. The test may take a minute or two.

Now, you can see a screenshot of the page exactly how Google renders it. To check whether the search engine is reading the code correctly.

Just click the “View Tested Page” link and then the “Screenshot” tab.

"View Tested Page" clicked on the left and the "Screenshot" tab clicked on the right of the "URL Inspection" page on GSC

Check for discrepancies and missing content to find out if anything is blocked, has an error, or times out.

Our JavaScript SEO guide can help you diagnose and fix JavaScript-specific problems.

Structured Data Issues

Structured data is data organized in a specific code format (markup) that provides search engines with additional information about your content.

One of the most popular shared collections of markup language among web developers is Schema.org.

Schema helps search engines index and categorize pages correctly. And help you capture SERP features (also known as rich results).

SERP features are special types of search results that stand out from the rest of the results due to their different formats. Examples include the following: 

  • Featured snippets
  • Reviews
  • FAQs
the featured snippet for the term "benefits of pizza" highlighted on the SERP

Use Google’s Rich Results Test tool to check whether your page is eligible for rich results.

Google’s Rich Results Test tool-start with an input box to enter and test a URL

Enter your URL to see all structured data items detected on your page.

For example, this blog post uses “Articles” and “Breadcrumbs” structured data. 

Rich Results Test showing a blog post using structured data like “Articles” and “Breadcrumbs”

The tool will list any issues next to specific structured data items, along with links on how to address them. 

Or use the “Markup” thematic report in the Site Audit tool to identify structured data issues.

Just click “View details” in the “Markup” box in your audit overview.

"Markup" under "Thematic Reports" on Site Audit clicked

The report will provide an overview of all the structured data types your site uses. And a list of any invalid items.

"Markup" report showing metrics and graphs for pages with markup, pages by markup type, structured data by pages, etc.

Invalid structured data occurs when your markup doesn’t follow Google’s guidelines. This can prevent your content from appearing in rich results.

Click on any item to see the pages affected.

"Structured Data Items" on the "Markup" report with the "Invalid" column highlighted

Once you identify the pages with invalid structured data, use a validation tool like Google’s Rich Results Test to fix any errors.

Further reading: Learn more about the “Markup” report and how to generate schema markup for your pages.

8. Check for and Fix HTTPS Issues

Your website should be using an HTTPS protocol (as opposed to HTTP, which is not encrypted).

This means your site runs on a secure server using an SSL certificate from a third-party vendor.

It confirms the site is legitimate and builds trust with users by showing a padlock next to the URL in the web browser:

the padlock icon highlighted in the URL bar for a site using the HTTPS protocol

HTTPS is a confirmed Google ranking signal

Implementing HTTPS is not difficult. But it can bring about some issues. Here’s how to address HTTPS issues during your technical SEO audit: 

Open the “HTTPS” report in the Site Audit overview: 

"HTTPS" under "Thematic Reports" on Site Audit clicked

Here, you’ll find a list of all issues connected to HTTPS. And advice on how to fix them. 

HTTPS report on Site Audit with "Why and how to fix it" under "8 subdomains don't support HSTS clicked

Common issues include:

  • Expired certificate: Your security certificate needs to be renewed
  • Old security protocol version: Your website is running an old SSL or TLS (Transport Layer Security) protocol
  • No server name indication: Lets you know if your server supports SNI (Server Name Indication). Which allows you to host multiple certificates at the same IP address to improve security
  • Mixed content: Determines if your site contains any unsecure content, which can trigger a “not secure” warning in browsers

9. Find and Fix Problematic Status Codes

HTTP status codes indicate a website server’s response to the browser’s request to load a page. 

1XX statuses are informational. And 2XX statuses report a successful request. Don’t worry about these. 

Let’s review the other three categories—3XX, 4XX, and 5XX statuses. And how to deal with them. 

Open the “Issues” tab in Site Audit and select the “HTTP Status” category in the top filter.

"HTTP Status" selected as the category on Site Audit Issues showing a list of related issues

To see all the HTTP status issues and warnings.

Click a specific issue to see the affected pages. 

3XX Status Codes

3XX status codes indicate redirects—instances when users and search engine crawlers land on a page but are redirected to a new page.

Pages with 3XX status codes are not always problematic. However, you should always ensure they are used correctly to avoid any possible problems.

The Site Audit tool will detect all your redirects and flag any related issues.

The two most common redirect issues are as follows:

  1. Redirect chains: When multiple redirects exist between the original and final URL
  2. Redirect loops: When the original URL redirects to a second URL that redirects back to the original

Audit your redirects and follow the instructions provided within Site Audit to fix any errors.

4XX Status Codes

4XX errors indicate that a requested page can’t be accessed. The most common 4XX error is the 404 error: Page not found. 

If Site Audit finds pages with a 4XX status, remove all the internal links pointing to those pages.

First, open the specific issue by clicking on the corresponding number of pages with errors.

"365 pages returned 4XX status code" highlighted and clicked on Site Audit Errors

You’ll see a list of all affected URLs.

a list of page URLs that have returned 4XX status codes plus the **** they were discovered on Site Audit Issues

Click “View broken links” in each line to see internal links that point to the 4XX pages listed in the report. 

Remove the internal links pointing to the 4XX pages. Or replace the links with relevant alternatives. 

5XX Status Codes

5XX errors are on the server side. They indicate that the server could not perform the request. These errors can happen for many reasons. 

Such as:

  • The server being temporarily down or unavailable
  • Incorrect server configuration 
  • Server overload

Investigate why these errors occurred and fix them if possible. Check your server logs, review recent changes to your server configuration, and monitor your server’s performance metrics.

10. Perform Log File Analysis

Your website’s log file records information about every user and bot that visits your site.

Log file analysis helps you look at your website from a web crawler’s point of view. To understand what happens when a search engine crawls your site.

It’s impractical to analyze the log file manually. Instead, use Semrush’s Log File Analyzer.

You’ll need a copy of your access log file to begin your analysis. Access it on your server’s file manager in the control panel or via an FTP (FileTransfer Protocol) client

Then, upload the file to the tool and start the analysis. The tool will analyze Googlebot activity on your site and provide a report. That looks like this: 

"Log File Analyzer" with different charts showing Googlebot Activity, Status Code, and File Type

It can help you answer several questions about your website, including:

  • Are errors preventing my website from being crawled fully?
  • Which pages are crawled the most?
  • Which pages are not being crawled?
  • Do structural issues affect the accessibility of some pages?
  • How efficiently is my crawl budget being spent?

These answers fuel your SEO strategy and help you resolve issues with the indexing or crawling of your webpages.

For example, if Log File Analyzer identifies errors that prevent Googlebot from fully crawling your website, you or a developer can work to resolve them.

To learn more about the tool, read our Log File Analyzer guide.

Boost Your Website’s Rankings with a Technical SEO Audit

A thorough technical SEO audit can positively affect your website’s organic search ranking.

Now you know how to conduct a technical SEO audit, all you have to do is get started.

Use our Site Audit tool to identify and fix issues. And watch your performance improve over time. 

This post was updated in 2024. Excerpts from the original article by A.J. Ghergich may remain.





Source link : Semrush.com

Social media & sharing icons powered by UltimatelySocial
error

Enjoy Our Website? Please share :) Thank you!