Data Anomalies In Search Console - Page Actions Are Temporarily Disabled

Data Anomalies In Search Console - Page Actions Are Temporarily Disabled

Google Index Update on twitter

When Google launched a new search console in 2018, it promised users of coming up with a URL inspection tool. This tool helps to optimize and monitor the website performance. With the help of the tool, users can easily see the Google search status of specific links. On October 14, 2020, Google temporarily drops request indexing features. After Google announced about URL Inspection tool update on Twitter, it is now available for users. Data anomalies in Search Console Changes to stay updated with current trends. This update lets users check the URL of the website and understand how Google search crawls the URL.

Data anomalies in Search Console

What Does The URL Inspection Tool Do?

The URL inspection tool allows users to enter a link and fetch details about its crawling, indexing, and other related information.


Crawling is the process in which Google sends out a team of spiders or crawlers to find new and updated content on a website. The program that fetches the content is called Google bot. Google bot uses a certain algorithm to find out which sites to crawl, at what specific intervals of time, and how many pages to fetch from each site. Artificial Bot can scan any type of content including a web page, an image, a video, a PDF, etc.

This crawling process starts with a list of web page URLs, resulting from previous crawl processes. When a Googlebot visits a page, it discovers links on the page and includes them to its list of pages to crawl.

Any changes to existing sites, new sites, and dead links are identified and used to update the Google index. During the crawling process, Google renders the page using the newest version of chrome. In the rendering process, it runs any page scripts it finds.

Primary Crawl/Secondary Crawl

Google search engine uses both mobile and desktop crawler to crawl the website. The primary crawler crawls all the pages on your site. Along with that, Google recrawls a few pages on your site with a secondary crawler. This will help you to see how well your site works with the other device type.

How Does Google Identify Which Pages Not To Crawl?

Google does not crawl the pages that are blocked in robots.txt but it can be indexed if linked to by another web page.

It also doesn’t crawl any pages that are not accessible by an anonymous user. For example, any page that asks for login information to get inside will not be crawled by Bots.

Duplicate pages are crawled less frequently.

Steps To Boost Your Crawling

Follow the steps to improve your crawling by Google

  • Submit sitemap and crawl requests for individual pages

  • Use readable URLs and include direct internal links within the site

  • Make use of robots.txt wisely

  • Add hreflang to denote another version of your pages in various languages

  • Identify your alternate and canonical pages 

  • Using the index coverage report, view your crawl and index coverage

  • Make sure that Google can access images, CSS files, and Scripts


After crawling the Page, Google will process the page to know the inner content of the page. It processes everything on the webpage including textual content, key content tags, attributes, alt attributes, images, videos, and more. But, the bots cannot process rich media files such as HD graphics, high-quality images, and more.

Google easily finds out if a page is duplicate or not within a short period. If a page is identified as duplicate or canonical, then it will be crawled much less frequently. We have to know that Google doesn’t index pages with a noindex directive. It scans the page even though it is blocked by a robots.txt file, a login page, or another device.

Improve Your Indexing

There are many ways to improve google’s ability to understand the exact content of your page

  • Don’t try to ‘index’ a page that is being blocked by robots.txt. If you do so, the index is not noticeable and the page might still be indexed.

  • Use structure data to quickly index your page

  • Stick with Google webmaster guidelines 

What Is A Document?

Google denotes the web as a set of several documents kept in a single platform. Each document represents one or more web pages. These web pages are almost similar but are essentially the same content, easily reachable by different URLs. The various URLs in a document can direct you to the exact same page. Google selects one of the URLs in a document and makes it as the document’s canonical URL. The canonical URL is the most crawlable and indexable in the document. The other URLs are considered duplicates and it is crawled occasionally by bots. For example, in particular, websites, if a document’s canonical URL is the mobile URL, then Google will still probably serve the desktop URL for users searching on the desktop, reports in Search console attribute data to the website document’s canonical URL. URL inspection tools support testing alternate URLs, but inspecting the canonical URL should deliver data or information about the alternate URLs as well.

Serving Results

When a user enters a query, Google searches the index for matching pages and return the results we believe are the most relevant to the user.


This URL Inspection tool update gives us a detailed report about the crawl, index, and information essential for analytics. It gives information related to indexing errors, crawl errors, and the status of the previous crawl.