This checklist is designed for all kinds of SEOs who aspire to improve the website’s visibility, increase site’s indexation, and boost organic traffic at the end of the day. This includes accessibility for proper indexation, optimized site structure, beneficial interlinking, elimination of technical errors, and other aspects of on-site SEO. You have your hands on an easy-to-follow action plan with the help of which you will definitely know What to change so to get high rankings in SERP.
Log files provide SEOs with invaluable data. It’s extremely handy for your daily SEO routine as well as for more experimental tasks. It reveals not only technical issues and showcases your crawl capacity, but also provides valuable SEO insights. Most importantly, the data obtained from log file analysis provides a bulletproof argument that supports your SEO strategy and experiments. When you have facts to support your theory, it becomes reasonable in web developers’ eyes (especially when their agenda doesn’t initially align with yours).
Furthermore, log files could be helpful for conversion optimization.
We`ve already mentioned about it in our article Logs Analysis For SEO Boost.
Caution! Log analysis opens the black box. You’ll be amazed by the secrets your website holds.
Crawl Budget is the number of pages Googlebot crawls and indexes on a website within a given period.
In other words, the crawl budget is the number of URLs Googlebot can and wants to scan. Yes, the crawl budget is limited and defined for each website individually. If your website has more pages than it’s set for crawling, Google will miss these extra pages. This, in turn, will affect your site’s performance in search results.
According to Google, several major factors exacerbate a site’s crawlability and indexability. Here are the factors in order of significance:
1. Faceted navigation and session identifiers
2. On-site duplicate content
3. Soft error pages
4. Hacked pages
5. Infinite spaces and proxies
6. Low quality and spam content
Wasting server resources on pages like these will drain crawl activity from high-value pages, significantly delaying the in discoverability of great content on the site.
But the good news is that the crawl budget can be extended.
The very first step in optimizing your Crawl Budget is checking the site’s crawlability. Let’s start with a quick overview. The ‘Log Overview’ tab provides you with the general crawl data:
Click on Bot Dynamics tab for a more detailed review:
This dashboard showcases your crawl budget for each search bot. It can be extremely useful to quickly identify anomalies. You should check for deviations like spikes or declines.
However, the insights derived from this dashboard cannot be viewed independently. So, it’s time to go deeper for an eye-opening experience.
Crawl budget waste is a significant issue. Instead of valuable and profitable pages Googlebot oftentimes crawls irrelevant and outdated pages.
Go to the ‘Impact’ section to evaluate the Crawl Ratio and missed pages:
Once the crawl budget waste is obvious, it’s time for action. Each column is clickable and leads to the list of problematic URLs. JetOctopus crawl data provides you with the insights what problems there are on these pages and therefore what to change so to get all your business pages visited by search bots.
Orphaned pages are the pages that have no inlinks pointing to them. They can be irrelevant, for example, inherent to previous site structures. Also, these pages can be valuable and profitable.
First and foremost, exclude irrelevant orphaned pages from crawling and indexation:
If these pages are orphaned by mistake, you should include them in the site structure. This would improve their performance.
Crawl budget waste inevitably results in some pages being not visited by Googlebot.
These pages could be ignored for a myriad of reasons, the most common ones being:
A closer look to crawl data will help you improve their crawlability, indexability, and rankings.
Read the Case Study to see how TemplateMonster has handled both (orphan and not visited) types of pages.
Distance from index (DFI) is equal to the clicks from the homepage for each page on the website. It’s an essential aspect that impacts the frequency of Googlebot’s visits.
You can access the dashboard from the ‘Impact’ tab.
The following chart demonstrates that if DFI is deeper than 4, Googlebot crawls only half of the web pages or even less. The percentage of processed pages reduces if the page is deeper:
Indeed, distance from the homepage is very important. Ideally, high-value pages should be 2-3 clicks away from your homepage. Anything that is 4 and more clicks away will be seen by Googlebot as less important.
Check if these not-visited pages are valuable and profitable. If yes, improve their DFI by linking them from pages with smaller DFI.
Googlebot prioritizes pages that have lots of external and internal links pointing to them. This means that the number of inlinks is a significant factor impacting crawlability and rankings.
With JetOctopus data you can optimize your interlinking structure and empower your most valuable pages.
You can find this chart on the Impact page:
As you can see, there’s a correlation between the number of inlinks and crawled pages.
Firstly, check whether these not-visited pages are high-quality and profitable. If yes, increase the number of inlinks, especially from relevant pages with better crawl frequency.
Content plays a significant role in SEO. Google’s algorithms are constantly evolving to deliver “…useful and relevant results in a fraction of a second.” The more relevant and engaging your content, the better your site’s performance.
This chat from the Impact tab demonstrates the correlation between content size and bot visits:
Pages with less than 500 words have the worst crawl ratio, and often considered as low- quality pages.
Firstly, decide whether you need these pages. If yes, enhance the page content using Google’s Guidelines.
Titles are critical to giving both users and search bots a quick insight into the content of a result and its relevance to the keyword. High-quality titles on your web pages ensure a healthy crawl rate, indexation, and rankings.
There are two major crawlability issues associated with the title tags:
These errors most likely cause insufficient crawlability. Check this chart on Bot Behaviour by Title Duplications:
Firstly, decide whether you need these pages with empty or duplicate titles. Optimize title tags using Google’s Guidelines.
“URLs that are more popular on the Internet tend to be crawled more often to keep them fresher in our index.” – Google.
Pages that are most visited by Googlebot are considered as the most important ones. You should pay the closest attention to such URLs, keeping them evergreen and accessible. You can find the most visited URLs by clicking on charts in the ‘Pages by Bot Visits’ report:
You can link from these pages to the weaker ones (but still relevant) to improve their crawlability and rankings.
By clicking on ‘Pages by Bot Visits’ you can evaluate least visited pages.
Even though the frequency of Googlebot’s visits doesn’t necessarily correlate with good rankings, it’s a good practice to check high-value pages among these pages with a couple of visits. If there are some of profitable pages, you can analyze our data and devise a data-driven strategy to improve these pages overall performance:
Googlebot visits non-indexable pages, such as:
To ensure Googlebot isn’t visiting non-indexable pages, go to the Impact Dashboard:
The most common reason for such pages being crawled is the internal links pointing to them. Or maybe, some other page points to the non-indexable as a canonical or alternative lang version (hreflang tags). This, in turn, causes crawl budget waste.
Check and do your best so the non-indexable pages:
For more details: Wanna know which pages bots visit/ignore and why? Use logs analyzer
HTTP status codes have the biggest impact on SEO. Make sure your crawl budget is not spent on 3xx-4xx pages, and keep an eye on 5xx errors.
You can find the breakdown of status codes on Bot Dynamics dashboard:
We will discuss what you should do with:
500 server error and server’s response time are crucial because they affect access to your site. Search engine bots and human visitors alike will be lost. A significant number of 5xx errors or connection timeouts slows down the crawling.
You can quickly spot critical issues via the Health and Bot Dynamics Dashboards. If something goes wrong for Googlebot, you’ll be able to notice it in real-time and act accordingly.
Investigate and fix those issues as soon as you encounter them. If those errors are frequent, consider optimizing your server’s capacity.
3xx status codes are used for redirection. The most common are permanent redirects (301 status code) and temporary redirects (302 and 307 status codes).
Bot Dynamics and Health dashboards contain such data:
Investigate 3XXs. Why are they being crawled?
The most common 4XXs are:
Google spends crawl budget on these pages:
Investigate 4XXs. Why are they being crawled?
If the pages returning 404 codes are high-authority pages with lots of traffic or Googlebot’s visits, you should employ 301 redirects to the most relevant page possible.
Site speed significantly affects your crawl budget. For Googlebot a speedy site is a sign of healthy servers, so it can get more content over the same number of connections.
Check HTML load time for search bots and its dynamics on Bots Dynamic dashboard:
Improving the site speed is a rather complex task. The issue could be solved only by your web development team.
According to Google, they will be switching to mobile-first indexing for all websites, starting September 2020. In the meantime, they will continue moving sites to mobile-first indexing when their systems recognize that the site is ready.
If your website has not been switched to mobile-first indexing yet, keep an eye on Google’s mobile bot and its behavior.
Make sure your website is mobile-friendly. Once more, check Google’s Guidelines.
Log files also showcase how organic visitors perceive your site and the technical issues they are facing. This section is especially important for optimizing conversions.Evaluate how organic visitors explore your website. Refer to this bar chart.
1. Evaluate how organic visitors explore your website. Refer to this bar chart.
2. Identify technical issues encountered by your visitors:
By resolving these errors your site’s user experience will improve, thus increasing your conversion rate.
Overall, log file analysis is a must-have tool to boost your site’s crawlability, indexability, and rankings.
In combination with crawl data and Google Search Console data, log files analysis will give you powerful insights to increase your organic traffic and conversions.
With JetOctopus Logs tool you’d be able to:
Choose the way of log file integration: 3 ways of log file integration with JetOctopus.
Unlike the standard web crawlers and tools, Log files will tell you the full story. They can tell you how search engines are behaving on your site. Which pages they frequent, which pages they encounter problems, how fast your server is responding and many other essential things. Now, you take this information and overlay on top of the web crawler data, and it will open up a whole different world. What if you can verify errors reported by a Web crawler? How about overlaying your website conversion data on top of the crawled URLs? How about seeing how a specific page is crawled over time by search engines as it goes through various updates? How about finding orphaned pages? You can do this a lot more accurately than using conventional methods. There are many use cases for log file analysis and it’s often more powerful when you combine it with a Web crawler.
Log File analysis is a good way to better understand Google’s algorithms, including which pages are important (with intensive crawling and hundreds of hits per day) and which are not (with minimal crawling activity). It is also a great way to receive instant feedback from Google since when we analyze SERP positions, there is a huge delay between what was done and seeing the results of such. The biggest win for me has been optimizing my site’s structure after analyzing orphan pages and crawl depth, which has led to significant results.
It doesn’t matter how clever your internal linking system is – Google will always have its own view on how to crawl your website. Google behavior on different language versions of our website (which are almost identical) was surprisingly rather different.
Nowadays SEO requires you to push as many experiments as your dev team can. Even at the cost of extensive testing which sometimes leads to unexpected results.
With fast and highly customizable JetOctopus filtering and analytics tools, we were able to find errors and even whole problematic page clusters (50k+ 404 pages still under crawl).
Using the JetOctopus log analysis, we’ve reviewed our indexing policy. After about a year that helped us achieve 80–90% crawled pages by Google of all known, and the number of active pages increased by 60%.
Implementing improvements from JetOctopus reports on a regular basis, we have prevented a few (possibly big) indexing issues in our tests and now it is a vital part of our day-to-day routine.
Well, one thing that comes in my mind, hasn’t been covered that much and what many don’t necessarily understand is that the data in logs is exact. Not like in GA where blockers, the timing of the data push, sampling, etc. can affect the data. In most cases, logs are 100% accurate. So when you analyze the data it is trustworthy by default. Ofc it is also unfiltered by default so might need to clean it up a bit. Like internal traffic etc.
But from that data whatever analysis you do after proper filtering is pretty much something you can be confident to make decisions on.
The biggest challenge in log analysis – is to understand how all this data is connected to your website. What it all means. Without it, you can’t make the right conclusion, and all the insights can be inaccurate.
I am glad that I can analyze millions of log data lines overlaying them with crawl for 3 Mln urls and GSC data for a month. Because it gives you really interesting data joins: how many noindex pages are crawled by Googlebot; what are the most effective pages by clicks in logs and a lot of other questions can be answered with the help of logs overlapped with other datasets.
My last colossal finding was a segment of the pages which got bot’s visits over the given period of time. We’ve found millions of trash pages crawled by Googlebot. We closed all these pages in robots.txt and saved the crawl budget for the profitable pages. I know that it will lead to better indexability shortly.
Log analysis is one of the most important parts in the field of technical SEO. Don’t be afraid to jump into it! It allows you to have an exact and unlimited trace of the paths from robots. Some technical problems are only visible thanks to the logs, which makes the importance of an analysis primordial in a global SEO strategy.
The other great strength of the logs is to be able to cross this set of data with a crawl to easily and quickly identify orphaned URLs or hit pages that should not be. In constantly changing sites, the logs provide a historical trace of the evolution.
All this data allows you to demystify the notions of Crawl Budget and optimize the path of robots to important pages.
Bot activity check is my everyday task while I am drinking my morning coffee 🙂
Log analysis together with convenient data tables merge also gave me an opportunity to shape interlinking effectively and get rid of unnecessary pages. For example, we identified that googlebot has been knocking to our API and been making requests hundreds times per day. So, we optimized crawl budget and reduced the load on our servers 🙂
Doing SEO on the top level with large or huge websites without analyzing logs is just impossible.
For example, without logs on an e-commerce site you won’t be able to know:
which page type is most crawlable, which categories can cause problems, is your internal linking strategy correct, does your filter navigation work properly. Or some really basic stuff like if there are new problems after recent dev release? Are you dealing with pages with 5XX status? Do you properly handle 404 and 301 pages?
Also, keep in mind that with tools like JetOctopus you can combine your logs with GSC or crawl data. Imagine you can easily see the correlation between content quality and crawl rate or from a business value perspective – number of products on a specific category and crawl rate. By combining logs with custom extraction, the only thing that is limiting you is your own imagination.
Elena Bespalova, SEO Strategist at Rank Media Agency
Log analysis has become a part of my SEO-routine. It’s really helpful not only for the regular health checks but also for more advanced tasks.
Log analysis showcases weak spots and provides useful insights. To me, the greatest challenge was a migration of a storied website with 300k+ indexed URLs and even more URLs inside the black box. With the log analysis, we discovered tons of orphaned pages, inconsistency in interlinking structure, and technical issues.
With this data, the migration became less challenging. We could successfully map all the required redirects, optimize the site’s structure, prevent migration of technical
errors, and optimize the interlinking structure. Long story short: the migration proved to be successful and organic traffic increased.
Crawl budget is one of the main components of the effective website scanning. The garbage technical pages appear on websites quite often (especially on the old ones), and it’s difficult to identify those pages without proper analysis.
With the help of log analysis of our site, we have found tens of thousands of junk pages that should not have been visited by Googlebot. We have managed to significantly optimize the crawlability of important pages after we’ve worked with the website structure, tags, etc.
Log file analysis allowed us to identify many unacceptable server responses (503, 302 in particular), which were received by Googlebot in large numbers during the crawling process. These problems were solved on a server-side in a short time.
Having a lot of design templates on our site, the log analysis showed us a clear picture of the bots` behavior on pagination pages. Our goal is to index tens of thousands of such templates, hence analyzing logs is extremely insightful for our website.
Also, the log files allowed us to identify the pages most visited by bots. In one click, we received a huge amount of information and loaded the backlog for months ahead.
When I performed my first log analysis with JetOctopus… I was shocked! I realized that Googlebot keeps attending hundreds of thousands of outdated, junk, and “gone forever” pages, even out of Google index. They were consuming our website crawling budget dramatically. Another insight – now we know the minimum number of internal links that guarantee that your target page will be visited by Googlebot. And what’s exciting – there are dozens of new SEO insights ahead.