must know the whole journey Google Bot Mobile Google Servers Newly crawled pages are indexed & added to server if allowed Browser Algorithms act as filter and show relevant results in index User searches & request sent to Google server… Relevant results listed in browser
brand pages can lead to KPI success More Organic Traffic Better Site Authority More Organic Revenue Better User Experience More SERP Presence Advantages Over Competitors
Robots.txt Issues Robots Directive Issues Internal Link Issues Technical Issues Are The Main Culprits The following SEO issues hinder a website's crawlability and indexability
working pages in sitemap(s) Do: ➔ Include important pages like category & product. ➔ Have < 50,000 URLs per sitemap. ➔ Include 200 code indexed pages. ➔ Group sitemap(s) by category, language, site area, etc. Don't: ➔ Include low value pages (T&C’s, etc.) ➔ Have > 50,000 URLs per sitemap ➔ Include non-indexed pages (e.g. canonicalised, redirected, noindex tag, etc.)
with incorrect tags <meta name = “robots” content=”noindex, nofollow”> Directives Directives Key: Index - Allows page to be indexed. Follow - Allows links on page to be crawled. NoFollow - Tells crawlers not to crawl links on the page. NoIndex - Tells crawlers not to index the page.
crawling of any .php$ resources and also disallowing crawling of any .php$ resources. Correct types of pages to disallow from Googlebot(s). Tidy Your Robots.txt Instruction File Check we’re not disallowing any important pages or areas
any rendering/speed/size issues. Product images are pulled from different CDN, which doesn’t auto compress. Use CDN & CMS Auto Compression Features Next.js (REACT Framework to create web applications) isn’t autocompressing using gzip, brotli or deflate. To combat Core Web Vital (CWV) issues such as LCP
increase Googles chances of caching and crawling our navigation links Googlebot mobile struggling to access links in menu. SSR will improve chances of caching and speed up the process in which content is loaded/rendered.
rendered first. Render Important Page Resources First This will allow Google to quickly find important page content Sometimes this isn’t possible, so SSR would help preload resources above images (e.g. JS)
can be used for indexation checks Shows us if the URL is indexed and in Google… Shows us what URL’s on the domain are indexed and ranking for the term “Hello World”… Shows us if the page is indexed and in Google for the term “Hello World”…
we identify how visible URLs are 1 2 3 Navigation is heavily JavaScript and hasn’t been cached by Google. This issue leads to less navigation URLs becoming less visible, impacting crawling & indexation.
before we dive into results END. Clothing is a global fashion retailer selling designer brands END. Clothing is JavaScript heavy & struggles with being crawled Tug identified crawling blocker as reason for lack of indexability
increased by 64% MoM Validation passed: Google resolved canonical issue There was a significant fix on 29th November, resolving a duplicate issue where there was a large volume of pages without user-selected canonical.