Why is my website not indexed?

12/10/2022
2192
0
Why is my website not indexed?
Why is my website not indexed?

To begin with, you need to check whether your site or its individual pages are represented in the Google index. Most often, there is no problem, just that the page does not rank high enough in the search results.

Effective way to check:

  1. Turn off safe search, or go to the anonymous browser tab
  2. Find the page you are interested in using the site:your-domain search operator – you can read more about search operators here

Examples: site:example.com or site:example.com/zootovary

  1. If you managed to find the required page, it means that it is in the index

What are the reasons for the absence of the site in the index and what to do to eliminate them

As for the options, why Google does not index the site as a whole, you need to carefully study each individual case, and find the source of this phenomenon as soon as possible.

The site is closed from indexing in Robots.txt

One of the most common mistakes webmasters make is incorrectly compiled robots. The presence of blocking Disallow directives in it can prevent the bot from processing the entire resource or its individual web pages. Here it is important to remember that every website has pages that need to be closed from indexing. These include technical search results, get-parameters, login, admin, shopping carts, “trash”, etc. Read more about configuring the Robots.txt file in our article «How to make a robots.txt file?» Ways to check the correctness of the Robots.txt file setting:

  • Add the ending /robots.txt to your domain name and check it yourself
  • Open to check the corresponding robot a tool from Google. But first you need to add the site to Google Search Console
  • Scan the site with Screaming Frog Seo Spider or Netpeak Spider. In this program, you will go to pages that are closed for indexing, correct errors in robots.txt and check the correctness of the functioning of the corrected file.
  • Use the search engine and find another service convenient for you to check the correctness of the J file setting

Robots meta tags

Sometimes the pages of the site can be closed using a tag that is set in the code of the pages. It looks like this <meta name=“robots” content=“noindex”> – prohibition of site content scanning. There are many directives, but let’s stop at this one. So, when this tag is installed on the page between the <head></head> tags, when Google scans the site, it does not add to the index the pages that have the robots tag with the noindex directive, so you need to check this moment when you notice that the site pages are not included in index. How to check Robots meta tags:

  • A tool for webmasters

In Google Search Console, open the Index – Coverage – Excluded tab

  • Using Screaming Frog Seo Spider or Netpeak Spider
  • Open the code of the page manually (F12), and by pressing the key combination Ctrl+F find meta name=”robots”

Prohibitions in the .htaccess file

The .htaccess file contains the rules of the web server. This file is usually placed in the root directory of the web server or the root directory of the site. With the help of some .htaccess rules, you can close the site from indexing. Check the .htaccess file on your server to see if it might contain any indexing rules.

Incorrectly configured or missing rel=”canonical” tag

The rel=”canonical” tag is used on pages with the same content. This tag tells search engines the address of the page that is the main page. Consider, for example, two pages that have the same content:

  • https://mybestua/original-page/ – main page;
  • https://mybestua/dublicat-page/ – a page with duplicate content.

To have the main page in the index, you can apply the rel = canonical tag. In the html code of the page https://mysupersite.com/dublicat-page/ between the <head></head> tags, you need to add the following tag: <link rel=”canonical” href=”https://mysupersite.com/original- page/” /> If your page or pages are not indexed, check the presence of the rel=”canonical” tag in the html code and its correctness.

Prohibition against X-assisted indexing is enabled‑Robots-Tag

X‑Robots-Tag is a directive in server response headers that can be used to prevent robots from indexing specific pages or files. Example of an HTTP response from a server with an X-Robots-Tag directive that prohibits page indexing: HTTP/1.1 200 OK Date: Tue, 11 May 2022 22:34:11 GMT (…) X-Robots-Tag: noindex (…)

Other reasons

  • Google Search Console does not list all domain variants (http:// and https://)
  • The site is sanctioned by the search engine
  • The site is not adapted for mobile devices
  • Technical SEO issues
  • Low response speed of the website server
  • The site is too young, or the domain has a bad history

To summarize

For a website to be successful, it is important to put good content on it, work with technical SEO and create quality backlinks. But all this will be useless if it is not indexed. So, make sure all indexing issues are resolved and then Google will thank you with good traffic.

Stanislav Nikitiuk
Linkbuilder
commercial offer

    SEO promotionCopywritingSMM promotionDevelopmentContextual advertisingDesign
    Digital новини в нашому телеграм-каналі
    Інтернет-маркетинг
    простою мовою
    subscribe
    Other articles by the author
    20/07/2021
    There are many different services for checking website traffic. Here we will consider the most popular ones, namely: SimilarWeb; Ahrefs; Serpstat; SemRush.

    12/10/2022
    For a website to be successful, it is important to post good content on it, work with technical SEO, and build quality backlinks. But all this will be useless if it is not indexed. Therefore, make sure that all indexing problems are solved, and then Google will thank you with good traffic.

    05/06/2024
    There are certain recommendations to get your business to the top of search results (even Wikihow says so). And we have a proven formula on how to do it. There are several opportunities to achieve higher positions on Google, as well as to be ranked first.

    Latest articles by #SEO
    24/04/2025
    Usually, SEO specialists use various services such as SEMRush, Serpstat, or Ahrefs to analyze various kinds of data (meta tags, keys, etc.). But all the powerful functionality of these tools is not always appropriate for some small or everyday SEO tasks.

    23/04/2025
    Search query content in SEO is the meaning that a user puts into their query to a search engine. Otherwise, intent can be called the goal, intention, and in some ways even the user's pain.

    01/04/2025
    The iFrame tag is a useful tool for diversifying the content on a website. WEDEX will tell you how the tag affects SEO and what you should pay attention to.

    WhatsApp Telegram Viber Почати розмову