Technical website positioning is principally about ensuring search engines like google and yahoo can discover, perceive, and correctly index (save) your website, but it surely goes past that. Technical website positioning additionally includes making a seamless consumer expertise in your guests.
Technical website positioning is the inspiration on which your website positioning success depends. As a result of many of those actions have to happen earlier than your different website positioning techniques have an opportunity of driving outcomes.
When approaching technical website positioning for any web site, following a structured course of is useful. That’s why we’ve put collectively this website positioning’s technical guidelines.

Whereas creating this guidelines, we additionally analyzed over 50,000 domains to see how widespread a few of these points are. We’ll share particular knowledge factors from our analysis as we go.
Now, let’s take a look at the guidelines gadgets, one after the other.
1. Search for Crawling and Indexing Points
Be sure search engines like google and yahoo like Google are in a position to uncover (crawl) and save (index) your website correctly to make sure your pages can rank in search outcomes.
Test Whether or not Your Web site Is Listed
This is absolutely the very first thing you’ll want to do as a result of your website received’t present in search outcomes if it isn’t listed.
The easiest way to verify your indexing standing is thru Google Search Console (GSC).
In GSC, head to the “Pages” report. This may present you precisely which pages are listed and that are excluded.

The pages that aren’t listed might be grouped by the particular cause.

Listed below are just a few causes you may see:
- Crawled – at the moment not listed: Google really checked out these pages however determined they weren’t value indexing. This often means the content material is low high quality or the pages are too just like current pages.
- Blocked by robots.txt: Your robots.txt file is telling Google to not crawl these pages. Double-check your robots.txt file to ensure you’re not unintentionally blocking vital content material.
- Excluded by ‘noindex’ tag: You might need unintentionally added noindex tags to your pages. This explicitly tells Google, “Do not put these pages in search outcomes.” Take away this tag from vital pages you wish to be listed.
Repair these points, then use the “Validate Repair” button that corresponds to a given concern to ask Google to recrawl your pages.

Test for Any Duplicates of Your Web site
Having duplicate variations of your website can hurt your website positioning efforts as a result of search engines like google and yahoo view these as separate web sites, regardless that they show the identical content material. As a result of they’re all technically totally different variations.
How typically do websites run into this concern?
In our research of over 50,000 domains, we discovered that 27% of internet sites had each HTTP and HTTPS variations accessible on the identical time. Which implies this concern is kind of widespread on the net.
For instance, your web site could be accessible at:
- https://yourdomain.com
- https://www.yourdomain.com
- http://yourdomain.com
- http://www.yourdomain.com
Test in case your website is accessible via a number of URLs by getting into every variation in your browser.
If a number of variations load, you’ll want to choose one most popular model and redirect all others to it.
Use the HTTPS model as your main URL (both with or with out www—that is your choice). Then implement a 301 everlasting redirect, so customers and search engines like google and yahoo are forwarded to your most popular model.
Make Positive Your Robots.txt File Is Precisely Set Up
An incorrectly configured robots.txt file can block vital pages from being crawled, so assessment your robots.txt file to verify it’s correctly configured.
From our dataset, solely 2% of internet sites had robots.txt configuration points. That’s not a variety of web sites, however the penalties may be extreme should you occur to have this concern.
A robots.txt file is a plain textual content file that tells search engine crawlers which elements of your website they’re allowed to entry—and which elements to keep away from.
The file incorporates strains that may appear to be this:
Consumer-agent: *
Disallow: /admin/
Disallow: /login/
Permit: /
Your robots.txt file might be positioned at “yourdomain.com/robots.txt.” Test the “Disallow” directives particularly to ensure you’re not blocking vital folders or pages.
Repair Redirect Chains & Loops
Redirect chains and loops harm your website positioning as a result of they sluggish issues down for customers, waste crawl price range (search engine assets used to crawl your website), and make it more durable for search engines like google and yahoo to cross full rating energy to your pages.
So, you need to verify for and repair these redirect points to maintain your website working easily.
A redirect chain happens when a URL redirects to a different URL after which to a different, as a substitute of straight reaching the ultimate vacation spot.

A loop occurs when a URL redirects to a URL that redirects again to the unique, creating an infinite loop.

Each of those points happen resulting from ongoing adjustments to your web site, corresponding to if you restructure URLs and don’t fastidiously handle redirects throughout the course of.
How widespread is that this downside?
About 12% of internet sites in our research had redirect chains or loops affecting their efficiency.
Use Semrush’s Web site Audit device to seek out redirect-related points in your website.
Arrange a challenge within the device to run a full audit. (It solely takes a couple of minutes for many websites.)
Then, go to the “Points” tab and seek for “redirect” within the search bar to see any redirect chains or loops.

And the particular URLs concerned:

Fixing these points is straightforward:
- For chains, replace any hyperlinks or redirects to level on to the ultimate vacation spot URL
- For loops, make certain URLs don’t redirect again to themselves
Repair Damaged Hyperlinks
Damaged inner hyperlinks direct customers to webpages that not exist, resulting in 404 error pages after they’re clicked. This leads to a poor consumer expertise and may sign to search engines like google and yahoo that your website is not well-maintained.

Damaged hyperlinks may be both inner hyperlinks to your personal content material or exterior hyperlinks pointing to different web sites.
In our evaluation of over 50,000 domains, we found that 52% of internet sites had damaged inner/exterior hyperlinks that wanted fixing.
The Web site Audit device may also help you simply spot if this concern exists in your website. Go to the “Points” tab and seek for “damaged.”
You’ll see errors if damaged hyperlinks are current.

When you click on on the detected gadgets, you’ll see the affected pages:

As soon as you have recognized damaged hyperlinks, you will have totally different choices relying on the kind:
For damaged inner hyperlinks:
- Restore the deleted web page if potential
- Arrange a 301 redirect to ship customers to an analogous related web page
For damaged exterior hyperlinks:
- Change the hyperlink with an up to date model of the web page if it exists elsewhere
- Take away the hyperlink fully if no appropriate alternative exists
- Discover another useful resource that gives related data and hyperlink to it as a substitute
It’s a good suggestion to schedule a damaged hyperlink verify each week. It is stunning how rapidly these can pile up, particularly on bigger websites.
You possibly can arrange automated crawls with Web site Audit.

Repair Server Errors
Server errors (5xx errors) are problematic as a result of they forestall search engines like google and yahoo from crawling and indexing your content material.
A server error signifies that there’s something flawed with the server itself. Our technical website positioning research revealed that about 10% of internet sites expertise some kind of server error regularly.
One of the crucial widespread server errors is the five hundred Inside Server Error, which is actually your server’s approach of claiming, “One thing went flawed, however I am unsure what.”

Different server errors embody:
- 502 Unhealthy Gateway: This happens when your server tries to speak with one other server to meet a request, however receives an invalid response
- 503 Service Unavailable: This usually happens when the server is overloaded with requests or is at the moment beneath upkeep
- 504 Gateway Timeout: This occurs when the server doesn’t reply in time
You possibly can verify for server errors by yourself by looking for “5xx” errors in Web site Audit.

When you see any errors, click on on the hyperlink to see the problematic pages and the particular error codes they’re returning. This data might be essential for troubleshooting.

Fixing server errors requires technical experience. So work along with your developer or internet hosting supplier to resolve them.
2. Optimize for Consumer Expertise
Engines like google are inclined to reward web sites that prioritize an excellent consumer expertise (UX). Plus, an excellent expertise retains guests engaged and encourages them to discover your content material.
Listed below are the principle features to deal with:
Make Positive Your Web site Is Cellular-Pleasant
Engines like google primarily use the cell model of your website for rating and indexing, so you need to make certain your web site shows and features correctly on smartphones.
To verify your website’s mobile-friendliness, use the Web site Grader device. Simply enter your area and e-mail deal with, and the device will analyze whether or not your website is optimized for cell.

The device will present you precisely the place your cell expertise falls quick and what wants fixing. So you’ll be able to work alongside your developer workforce to make the required adjustments.
Enhance Your Core Net Vitals
Enhance your Core Net Vitals. Google makes use of them as a rating issue.
The Core Net Vitals are a set of three most important metrics that consider your webpage’s general consumer expertise.
The precise metrics are:
- Largest Contentful Paint (LCP): Measures how rapidly the principle content material of your web page masses and turns into seen to customers. Google desires this to occur inside 2.5 seconds for an excellent consumer expertise.
- Interplay to Subsequent Paint (INP): Measures how rapidly your web page responds visually after a consumer interacts with it (like clicking a button or tapping a hyperlink). Particularly, it tracks the time from if you click on one thing to if you see the web page replace or change. This could occur in lower than 200 milliseconds for an excellent expertise.
- Cumulative Format Shift (CLS): Measures visible stability (i.e., how a lot parts leap round because the web page masses). Google desires your CLS rating beneath 0.1.
Not many web sites are totally optimized for these metrics.
In keeping with our research, 96% of internet sites had no less than one web page that failed the Core Net Vitals evaluation check.
You should utilize Google Search Console to see your Core Net Vitals efficiency.
In GSC, navigate to “Core Net Vitals” from the sidebar and click on “Open Report” to see the info.

Then, search for pages marked as “Poor” or “Wants Enchancment.” These pages have failed the Core Net Vitals evaluation check and wish optimization.

Take these URLs and run them via Google’s PageSpeed Insights device to get particular suggestions on how you can repair the problems.
Like this:

Work along with your developer to implement the advised fixes.
Keep away from Intrusive Interstitials
Keep away from intrusive interstitials as a result of they create a poor consumer expertise, particularly on cell gadgets with restricted display house.
Intrusive interstitials are pop-ups or overlays that cowl a good portion of your content material, making it troublesome for customers to entry the data they got here for.

What precisely counts as an intrusive interstitial? Listed below are just a few examples:
- Full display pop-ups that cowl the principle content material instantly after a consumer lands in your web page
- Stand-alone interstitials that customers should dismiss earlier than accessing your content material
- Layouts the place the above-the-fold space is occupied by an advert, however the principle content material is definitely beneath
Not all interstitials are problematic, although. Google makes exceptions for:
- Authorized obligations like cookie notices or age verification dialogs
- Login dialogs on websites the place content material is behind a paywall
- Small banners that use an inexpensive quantity of display house and may be simply dismissed
- Exit-intent pop-ups that seem when somebody is about to go away
- Small pop-ups that seem after a consumer has been on the web page for some time
3. Work On Your Web site Navigation
Be sure your website has a reasonably easy navigation system that can enable customers to seek out vital content material simply and assist search engines like google and yahoo perceive your website.
Enhance Your Web site Construction
A well-structured web site is essential for search engines like google and yahoo and customers as a result of it determines how effectively search engines like google and yahoo can crawl and perceive your content material and the way simply guests can navigate your website.
The best construction resembles a hierarchy that makes logical sense. Your homepage is on the high, adopted by most important class pages, then subcategories, and at last particular person pages.
This creates clear paths for each customers and search engines like google and yahoo to comply with.

Every web page must be accessible inside three or 4 clicks out of your homepage. As you’ll be able to see within the instance above.
Interlink Your Pages
Inside linking creates pathways between totally different pages in your website, permitting search engine crawlers to find your content material whereas serving to customers discover associated data.

Search for alternatives so as to add contextual hyperlinks inside your content material. When including hyperlinks, use descriptive anchor textual content fairly than generic “click on right here” or “learn extra” phrases.
The textual content you utilize for the hyperlink lets customers know the place they’re about to go and helps Google perceive what the linked web page is about.
Another efficient inner linking techniques embody:
- Linking out of your navigation menu to your most vital pages to present them additional authority
- Creating hub pages (most important subject pages) that carry collectively and hyperlink to all of your associated content material in a single place
- Including “associated posts” sections on the finish of articles
Do not go overboard, although. Too many inner hyperlinks on a web page can look spammy.
Use Breadcrumbs
Breadcrumbs enhance navigation and assist each customers and search engines like google and yahoo higher perceive your website’s construction.
They seem on the high of a web page and present the trail to that web page inside your website. Customers can click on on them to simply return to earlier sections.

Nonetheless, not all websites want to make use of breadcrumbs.
Breadcrumbs are solely value implementing in case your website is comparatively giant with a number of layers of hierarchical content material (like ecommerce shops).
Repair Orphan Pages
Repair orphan pages as a result of they’re troublesome for customers and search engines like google and yahoo to find—they haven’t any incoming inner hyperlinks.

These pages are inclined to accumulate on bigger websites fairly simply.
In our evaluation of over 50,000 domains, we discovered that 69% of these web sites had no less than one orphan web page. So, this concern is kind of widespread.
You possibly can verify in case your website has any orphan pages utilizing Semrush’s Web site Audit device. Go to the “Points” tab and seek for “orphan.”

Repair the difficulty by including hyperlinks to the orphan web page from different related pages.
4. Meet Different Technical Necessities
There are numerous different normal greatest practices that Google desires you to comply with in your website.
Use HTTPS
Hypertext switch protocol safe (HTTPS) is the safe model of HTTP that encrypts the connection between the consumer’s browser and your web site. To guard delicate data like login credentials, cost particulars, and different private knowledge.
Google has used HTTPS as a rating sign since 2014, giving a slight website positioning benefit to safe websites.
Extra importantly, trendy browsers mark non-HTTPS websites as “Not Safe,” which may erode consumer belief and improve bounce charges.

You possibly can implement HTTPS in your website by buying a Safe Sockets Layer (SSL) certificates. Many website hosting companies provide this if you join, typically without cost.
That is extra of a sophisticated technical website positioning process, so discuss to your growth workforce to get this applied if it’s not already.
Implement Hreflang for Worldwide Pages
Hreflang tags are used for multilingual or multiregional websites to inform search engines like google and yahoo which language and area every model of your web page is meant for, guaranteeing customers see essentially the most related model in search outcomes.
For instance, if we seek for the official Disney website within the U.S., we see the American English model:

If we do the identical in Germany, we see German model of the web page:

Implementing hreflang is straightforward. Simply add the suitable tags to the
part of every language/country-specific model of your web page.
For instance, in case your web site targets audiences in the USA, Germany, and Japan, add these hreflang tags to all three variations:
The primary tag signifies the default or fallback web page that must be proven to customers when no different variant is acceptable.
Different tags specify the totally different language or nation variations obtainable in your website, guaranteeing Google serves the suitable one primarily based on a consumer’s location and language settings.
Add Schema Markup
Schema markup is a sort of code that helps search engines like google and yahoo perceive the content material in your pages extra clearly.
Whereas it is not a direct rating issue, schema markup permits wealthy outcomes (particular listings on search outcomes pages), which may enhance click-through charges.
There are numerous varieties of schema markup, however you need to deal with essentially the most related ones in your particular content material varieties. Which can embody:
- Group (in your homepage)
- Product (for ecommerce product pages)
- Article (for weblog posts or information)
- Occasion (for occasion pages)
- Recipe (for meals websites)
- Overview (for services or products evaluations)
The simplest option to generate schema is to make use of a Schema Markup Generator.
To get began:
- Choose the kind of markup you want
- Fill out the shape you see on the left-hand facet
- The device will robotically generate the code for you

As soon as the code is generated, you’ll be able to add it to the
part of your web page’s HTML.
After implementation, use Google’s Wealthy Outcomes Take a look at device to confirm that your schema is applied accurately.

5. Resolve Content material-Particular Technical Points
This part of our tech website positioning guidelines addresses how you can resolve widespread content-specific technical points that may have an effect on your website positioning efficiency.
Tackle Duplicate Content material
Tackle duplicate content material points as a result of they’ll harm your website’s credibility and make it more durable for Google to index and rank your content material for related search phrases.
Duplicate content material is content material that’s similar or extremely just like content material that exists elsewhere in your web site or on one other web site.
About 41% of internet sites had inner duplicate content material points, based on our evaluation of over 50,000 domains.
Use Web site Audit to rapidly discover out you probably have duplicate content material by yourself website.
Simply seek for “Duplicate” beneath the “Points” tab.

Tackle duplicate content material points by:
- Including canonical tags to determine the first model of your content material
- Consolidating duplicate pages right into a single, most important web page and establishing 301 redirects from the previous URLs to the principle one
Repair Skinny Content material
Determine and enhance skinny content material as a result of it will possibly negatively impression your search engine rankings and consumer expertise.
Skinny content material is a time period search engines like google and yahoo use to explain pages that supply little to no worth to customers.
To determine skinny content material in your website, search for pages which are:
- Poorly written and don’t ship worthwhile data
- Copied from different websites
- Full of advertisements or spammy hyperlinks
- Auto-generated utilizing AI or a programmatic methodology
As soon as you have recognized skinny content material, take motion to enhance it:
First, take away or rewrite any straight copied content material.
That is essentially the most severe concern. Plagiarism is not simply an website positioning downside—it is an moral one, too.
For the content material that’s skinny however unique, beef it up with significant insights, examples, or sensible suggestions that really assist your viewers.
For pages overwhelmed with advertisements, cut back promotions and deal with the consumer expertise.
For auto-generated fluff, both rework it into one thing helpful or do away with it fully.
Guarantee All Pages Have Metadata
Metadata helps search engines like google and yahoo perceive your content material and match it to related search queries, and it is what customers see in search outcomes when deciding whether or not to click on in your web page.
So you need to guarantee all of your pages have full, optimized metadata to enhance your search visibility and click-through charges.
Metadata contains parts just like the title tag and meta description, which summarize the web page’s content material and function. And may be displayed in search outcomes.
(Technically, the title tag isn’t a meta tag from an HTML perspective. But it surely’s vital in your website positioning and price discussing alongside different metadata.)
Use Web site Audit to simply verify for various points along with your meta tags, like lacking meta descriptions or title tags throughout your total website.
Simply filter your outcomes for “Meta tags” from the “Points” tab. Click on the quantity subsequent to a difficulty for a full checklist of pages with that downside.

Then, undergo and repair every concern. To enhance your visibility (and presumably your look) in search outcomes.
From what we have seen, points with meta tags are fairly widespread throughout the net.
In our evaluation of greater than 50,000 web sites, we found {that a} staggering 70% had been lacking meta descriptions, and 10% did not also have a title tag on a few of their pages.
Put This Technical website positioning Audit Guidelines Into Motion
Now that you understand what technical website positioning motion gadgets to prioritize, it’s time to execute.
Semrush’s Web site Audit device helps you determine and repair all of the technical points rapidly and effectively.
Join a free account to get began immediately.