Google Search Console is an indispensable tool for SEO as it provides data about the organic performance of a website or page.
Understanding how users search for products and services, measuring your site performance in the search engines, and getting recommendations for improvements are vital to SEO.
Originally known as Google Webmaster Tools, Google Search Console is the SEO tool that most SEO professionals either do use, or should be using for insights and technical health.
Google Search Console, also known as GSC, is a free service from Google that allows site owners to monitor their overall site health and performance using data directly from Google.
Among its many facets, GSC provides several valuable reports, including:
GSC also allows site owners to take actions related to their site like:
Additionally, GSC regularly sends updates via email to verified owners and users indicating any crawl errors, accessibility issues, or performance problems.
While the data for GSC has been expanded from just 3 months to up to 16 months, it doesn’t start collecting data until you have verified your ownership of the property in question.
To get started with Google Search Console, you’ll need a working Google account – such as a Gmail account or an email account associated with a Google Workspace (formerly G Suite) for business – and you’ll need to be able to add code to your website or update the domain name servers with your hosting provider.
In this section, we’ll cover the following:
Since the data provided and the available processes in GSC would be quite valuable to your competitors, Google requires that site owners take one of several available steps to verify their ownership.
It’s important to pause here and talk about the two different types of properties you can verify in GSC: Domain and URL Prefix.
If you are verifying your domain for the first time with GSC, this is the property type you should select, as it will establish verification for all subdomains, SSL patterns (http:// or https://), and subfolders on your site.
There are two types of verification for this property type: TXT and CNAME. Both will require you or your site engineer to have access to change your site’s Domain Name System (DNS) records.
For TXT verifications (preferred):
For CNAME verifications:
Watch this video by Google for more details:
Once you have verified your domain, you can verify additional properties for this domain using the URL Prefix property type.
This verification method is used when you are unable to access your domain’s DNS records or when you want to verify specific URL paths under an existing Domain verification.
The URL Prefix verification allows you to verify:
Please note that this verification method will result in data that only follows the specified prefix.
While smaller websites may only need one verification, larger sites may want to individually track site health and metrics for subdomains and subdirectories to get a more complete set of data.
GSC gives you five options for verifying your site or sub-sections using this URL Prefix method:
These verification methods can take anywhere from a few minutes to a few days to replicate, so if you need to click the Verify Later button and return later, you can find that site or segment in the Not Verified section under your account’s properties.
Simply click on the unverified site and click Verify Later.
While Googlebot will most likely be able to find your site’s XML sitemap over time, you can expedite the process by adding your sitemaps directly into GSC.
To add a sitemap to GSC, follow these steps:
Note that you can add as many sitemaps as your site requires. Many sites will have separate sitemaps for videos, articles, product pages, and images.
The added benefit of including your sitemaps in this interface is that you can compare the number of pages your site has submitted to Google to the number of indexed pages.
To see this comparison, click on the three vertical dots to the right of your sitemap and select “See page indexing.”
The resulting page will display the number of indexed pages (in green) and pages that are “Not Indexed” (in gray), as well as a list of reasons those pages are not indexed.
It is vital that you control who has access to the data and functionality within GSC. Some functions, such as the Removals tools, can be very dangerous in the wrong hands.
The permissions settings associated with user types limit who can access these parts of GSC.
It is important to note that Google recently enhanced Search Console security by introducing a new feature for managing ownership tokens under Settings > Users and permissions > Unused ownership tokens
Tokens are simply those unique codes in HTML tags you set up in your head tag, HTML files you upload, or DNS TXT record values you set when verifying your website.
Now imagine a scenario where a website had multiple verified owners via HTML tag upload, and one of them left the company. If you remove that user from the search console, the issue is that the person can still regain access to the search console if you don’t remove their token from the unused ownership tokens page. This is an important feature, as website owners can now remove outdated verification tokens to prevent unauthorized access by former owners.
In Google Search Console, data is segmented into Dimensions and Metrics. You can use the data in these reports to measure your site’s performance, from page indexation to ranking and traffic.
In the Performance report, Dimensions group data into meaningful segments, such as Pages, Queries, Countries, and Devices. Metrics include data, such as Impressions and Clicks.
In the Pages report, a Dimension would include the reasons pages were not indexed, whereas Metrics would include the number of pages affected by that reason.
For Core Web Vitals, the Dimensions would be Poor, Needs Improvement, and Good. The Metrics would include the number of pages that fall into each category.
GSC is a valuable tool for SEO pros because it enables us to diagnose and evaluate pages from the perspective of Google. From crawlability to page experience, GSC offers a variety of tools for troubleshooting your site.
Long before a page can rank in the search engines, it needs to be crawled and then indexed. A page must be crawlable to be evaluated for search.
Whether you are experiencing crawl issues or not, it’s a good idea to regularly monitor the Crawl Stats report in GSC. This report tells you if it has encountered any issues with:
To use this report:
If there are any pages with 404 response codes, click on Not found to review these pages.
If your pages are not indexed, they can’t rank for your most important terms. There are a couple of ways to see which pages on your site are not being indexed in GSC.
The first place you may feel compelled to view would be under the Pages report, located in the left column of GSC.
While you can access a lot of information about pages on your site that are both indexed and not indexed, this report can be a bit misleading.
The Not Indexed section of the report includes pages you may have intentionally not indexed, such as tag pages on your blog or pages you want to reserve for users who are logged in.
The best report for determining the true extent of indexation issues your site is having is in the Sitemaps report.
To get there:
The resulting report is like the Pages report, but it focuses on pages your site has outlined as important enough to include in the sitemaps you have submitted to Google.
From here, you can review the Reasons columns in the “Why pages aren’t indexed” table.
For example, you may have several pages that have been Crawled but are currently not indexed. To evaluate one of the pages, do the following:
The HTML code displayed in the inspection tool reflects what Googlebot could crawl and render, which is especially crucial for JavaScript-based websites where the content does not initially exist within the static HTML but is loaded via JavaScript (via REST API or AJAX).
By examining the HTML, you can determine if Googlebot could properly see your content. If your content is missing, it means Google couldn’t crawl your webpage effectively, which could negatively impact your rankings.
By checking the “More Information” tab, you can identify if Googlebot could not load certain resources. For example, you may be blocking certain JavaScript files that are responsible for loading content via robots.txt.
If everything appears in order, click Request Indexing on the main URL inspection page. The field at the top of every page in GSC allows you to inspect any URL on your verified domain.
I would also recommend making sure Google doesn’t have issues when crawling robots.txt under Settings > robots.txt. If you have a robots.txt and Google isn’t able to fetch it ( for example, because of a firewall blocking access), then Google will temporarily stop crawling your website for 12 hours. If the issue is not fixed, it will behave as if there is no robots.txt file.
If you’re having difficulty getting pages into Google index or ranking in the search engines, it may be worth it to review the Core Web Vitals (CWV) across your site.
These are a measure of how your site performs for actual users, pulling in data from the Chrome User Experience Report (CrUX for short).
CWV measures three major usability metrics: Cumulative Layout Shift (CLS), Interaction to Next Paint (INP), and Largest Contentful Paint (LCP).
GSC provides a score for both mobile and desktop, and pages are lumped into Good, Needs Improvement, or Poor based on CWV scores.
To see the issues your pages may have:
Use these diagnostics to inform your developers, designers, and engineers, who can help you resolve those issues.
Tip: Make sure your website is using HTTPS since it improves your website security and is also a ranking factor. This is quite easy to implement because nowadays, almost all hosting providers provide free SSL certificates that can be installed with just one click.
If you are experiencing issues with indexation and ranking, it’s also possible that Google has encountered a security issue or has taken manual action against your site.
GSC provides reports on both Security and Manual Actions in the left column of the page.
If you have an issue with either, your indexation and ranking issues will not go away until you resolve them.
There are five ways you can use GSC for SEO in your daily activities.
The top part of the Search Console Performance Report provides multiple insights on how a site performs in search, including in search features like featured snippets.
Four search types can be explored in the Performance Report:
By default, the Search Console shows the Web search type.
Change which search type is displayed by clicking the Search Type button:
A useful feature is the ability to compare the performance of two search types within the graph.
Four metrics are prominently displayed at the top of the Performance Report:
By default, the Total Clicks and Total Impressions metrics are selected.
By clicking within the tabs dedicated to each metric, you can choose to see those metrics displayed on the bar chart.
Impressions are the number of times a website appears in the search results. If a user doesn’t have to click a link to see the URL, it counts as an impression.
Additionally, if a URL is ranked at the bottom of the page and the user doesn’t scroll to that section of the search results, it still counts as an impression.
High impressions are great because it means that Google is showing the site in the search results.
The impressions metric is made meaningful by the Clicks and the Average Position metrics.
The clicks metric shows how often users clicked from the search results to the website. A high number of clicks and a high number of impressions is good.
A low number of clicks and a high number of impressions is less good but not bad. It means that the site may need improvements to gain more traffic.
The clicks metric is more meaningful when considered with the Average CTR and Average Position metrics.
The Average CTR is a percentage representing how often users click from the search results to the website.
A low CTR means that something needs improvement in order to increase visits from the search results, whether it’s changing the page title or updating the meta description.
A higher CTR means the site is performing well with users.
This metric gains more meaning when considered together with the Average Position metric.
Average Position shows the average position in search results the website tends to appear in.
An average in positions of 1 to 10 is great.
An average position in the twenties (20 – 29) means that the result appears after the user scrolls through the search results, revealing more results. While this isn’t bad, it isn’t optimal. It can mean that the site needs additional work to boost it into the top 10.
Average positions lower than 30 could (generally) mean that the pages in question may benefit from significant improvements.
It could also indicate that the site ranks for a large number of keyword phrases that rank low and a few very good keywords that rank exceptionally high.
All four metrics (Impressions, Clicks, Average CTR, and Average Position), when viewed together, present a meaningful overview of how the website is performing.
The big takeaway about the Performance Report is that it is a starting point for quickly understanding website performance in search.
It’s like a mirror reflecting how well or poorly the site is performing.
The Search results report in the Performance section of GSC allows you to see the Queries and their average Position.
While most queries for which many companies rank in the top three in GSC are branded terms, the queries that fall in the five to 15 rank are considered “striking distance” terms.
You can prioritize these terms based on impressions and refresh your content to include those queries in the language of those pages.
To find these terms:
For many sites, Googlebot is very efficient and discovers new pages quickly. However, if you have a priority page that you need indexed in a hurry, you can use the URL Inspection tool to test the live URL and then request indexing.
To request indexing:
This process will add your URL to Googlebot’s priority queue for crawling and indexing; however, this process does not guarantee that Google will index the page.
If your page is not indexed after this, further investigation will be necessary.
Note: Each verified property in GSC is limited to 50 indexing requests daily.
GSC historical data are limited to a maximum of 16 months. Fortunately, you can overcome this issue by exporting GSC data into BigQuery, where it can be stored indefinitely. This allows you to access as much historical data as you desire. Since this action is not retroactive, you should start as soon as possible.
Read more: Google Search Console Data & BigQuery For Enhanced Analytics
While you can do many things within GSC, some real magic happens when you integrate the data within GSC into the many SEO tools and platforms available on the market.
Integrating GSC into these tools gives you a sharper view of your site’s performance and potential.
From desktop crawlers, like Screaming Frog and Sitebulb, to enterprise server-driven crawlers, like Lumar and Botify, integrating your GSC information can result in a more thorough audit of your pages, including crawlability, accessibility, and page experience factors.
Integrating GSC into large SEO tools, like Semrush and Ahrefs, can provide more thorough ranking information, content ideation, and link data.
Additionally, the convenience of having all your data in a single view cannot be overstated. We all have limited time, so these options can greatly streamline your workload.
Google Search Console has evolved over its nearly two-decade lifespan. Since its release as Google Webmaster Tools, it has provided valuable insights for SEO pros and site owners.
If you haven’t integrated it into your daily activities, you will quickly find how essential it is to make good decisions for your site optimizations going forward.
More resources:
Featured Image: BestForBest/Shutterstock
Most business owners get so caught up fighting fires in their business that they run…
Generative AI, SGE, and now AI Overviews have been hot topics since the launch of…
SEO metrics such as rankings are great, but what matters most is SEO’s impact on…
A new study by search industry expert Rand Fishkin has revealed that Google’s rollout of…
In a recent LinkedIn post, Gary Illyes, Analyst at Google, highlights lesser-known aspects of the…
Here is a recap of what happened in the search forums today, through the eyes…
This website uses cookies.
Leave a Comment