Professional Documents
Culture Documents
Here's What Happened When I Followed Googlebot For 3 Months
Here's What Happened When I Followed Googlebot For 3 Months
On internet forums and contentrelated Facebook groups, discussions often
break out about how Googlebot works – which we shall tenderly call GB
here – and what it can and cannot see, what kind of links it visits and how it
influences SEO.
In this article, I will present the results of my threemonthlong experiment.
https://outline.com/yWqNy6 1/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
Almost daily for the past three months, GB has been visiting me like a
friend dropping by for a beer.
Sometimes it was alone:
[// ::]: ... /page.html Mozilla/.
(compatible; Googlebot/.; +http://www.google.com/bot.html)
[// ::]: ... /page.html Mozilla/.
(compatible; Googlebot/.; +http://www.google.com/bot.html)
[// ::]: ... /page.html Mozilla/.
(compatible; Googlebot/.; +http://www.google.com/bot.html)
[// ::]: ... /page.html Mozilla/. (compatible;
Googlebot/.; +http://www.google.com/bot.html)
[// ::]: ... /page.html Mozilla/. (compatible;
Googlebot/.; +http://www.google.com/bot.html)
Sometimes it brought its buddies along:
[// ::]: ... /page.html Mozilla/. (X;
Linux x_) AppleWebKit/. (KHTML, like Gecko; Google Search
Console) Chrome/... Safari/.
[// ::]: ... /image.jpg GooglebotImage/.
[// ::]: ... /page.html Mozilla/. (Linux;
Android ..; Nexus X Build/MMBP) AppleWebKit/.
(KHTML, like Gecko) Chrome/... Mobile Safari/.
(compatible; Googlebot/ ; +http://www google com/bot html)
https://outline.com/yWqNy6 2/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
(compatible; Googlebot/.; +http://www.google.com/bot.html)
And we had lots of fun playing different games:
Catch: I observed how GB loves to run redirections and crawl images,
and run from canonicals.
Hideandseek: Googlebot was hiding in the hidden content (which, as its
parents claim, it does not tolerate and avoids)
Here’s what happened when I followed Googlebot for 3 months
Survival: I prepared traps and waited for it to spring them.
Obstacles: I placed obstacles with various levels of difficulty to see how my
little friend would deal with them.
As you can probably tell, I was not disappointed. We had tons of fun and we
became good friends. I believe our friendship has a bright future.
But let’s get to the point!
I built a website with meritsrelated content about an interstellar travel
agency offering flights to yetundiscovered planets in our galaxy and beyond.
e content seemed to have a lot of merits when in fact it was a load of
nonsense.
e structure of the experimental website looked like this:
https://outline.com/yWqNy6 3/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
I provided unique content and made sure that every anchor/title/alt, as well
as other coefficients, were globally unique (fake words). To make things
easier for the reader, in the description I will not use names like anchor
cutroicano matestito, but instead refer them as anchor, etc.
I suggest that you keep the above map open in a separate window as you
read this article.
e First Link Counts Rule says that on a page, Google Bot sees only the
first link to a subpage. If you have two links to the same subpage on one
page, the second one will be ignored, according to this rule. Google Bot will
ignore the anchor in the second and in every consecutive link while
calculating the page’s rank.
It is a problem widely overseen by many specialists, but one that is present
especially in online shops, where navigation menus significantly distort the
website’s structure.
In most stores, we have a static (visible in the page’s source) dropdown
https://outline.com/yWqNy6 4/15
30/11/2018
, ( p g )
Here’s what happened when I followed Googlebot for 3 months
p
menu, which gives, for example, four links to main categories and hidden
links to subcategories. During the mapping of a page’s structure, GB sees all
the links (on each page with a menu) which results in all the pages being of
equal importance during the mapping and their power (juice) is distributed
evenly, which looks roughly like this:
e most common but in my opinion, the wrong page structure.
e above example cannot be called a proper structure because all the
categories are linked from all the sites where there is a menu. erefore, both
the home page and all the categories and subcategories have an equal
number of incoming links, and the power of the entire web service flows
through them with equal force. Hence, the power of the home page (which
is usually the source of most of the power due to the number of incoming
links) is being divided into categories and subcategories, so each one of
them receives only percent of the power of the homepage.
How the structure should look:
https://outline.com/yWqNy6 5/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
If you need to fast test the structure of your page and crawl it like Google
does, Screaming Frog is a helpful tool.
In this example, the power of the homepage is divided into four and each of
the categories receives percent of the homepage’s power and distributes
part of it to the subcategories. is solution also provides a better chance of
internal linking. For instance, when you write an article on the shop’s blog
and want to link to one of the subcategories, GB will notice the link while
crawling the website. In the first case, it will not do it because of the First
Link Counts Rule. If the link to a subcategory was in the website’s menu,
then the one in the article will be ignored.
I started this SEO experiment with the following actions:
First, on the page.html, I included a link to a subpage page.html as a
classic dofollow link with an anchor: anchor.
Next, in the text on the same page, I included slightly modified references
to verify whether GB would be eager to crawl them.
To this end, I tested the following solutions:
To the web service’s homepage, I assigned one external dofollow link for a
phrase with a URL anchor (so any external linking of the homepage and
https://outline.com/yWqNy6 6/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
the subpages for given phrases was out of question) – it sped up the
indexing of the service.
I waited for page.html to start ranking for a phrase from the first
dofollow link (anchor) coming from page.html. is fake phrase, or
any other that I tested could not be found on the target page. I assumed
that if other links would work, then page.html would also rank in the
search results for other phrases from other links. It took around days.
And then I was able to make the first important conclusion.
Even a website, where a keyword is neither in the content, nor in the meta
title, but is linked with a researched anchor, can easily rank in the search
results higher than a website which contains this word but is not linked to a
keyword.
Moreover, the homepage (page.html), which contained the researched
phrase, was the strongest page in the web service (linked from percent of
the subpages) and still, it ranked lower on the researched phrase than the
subpage (page.html) linked to the researched phrase.
Below, I present four types of links I have tested, all of which come after the
first dofollow link leading to page.html.
e first of the additional links coming in the code behind the dofollow link
was a link with an anchor (a hashtag). I wanted to see whether GB would go
through the link and also index page.html under the phrase anchor,
https://outline.com/yWqNy6 7/15
30/11/2018 g p g p
Here’s what happened when I followed Googlebot for 3 months
despite the fact that the link leads to that page (page.html) but the URL
being changed to page.htmltesthash uses anchor.
Unfortunately, GB never wanted to remember that connection and it did
not direct the power to the subpage page.html for that phrase. As a result,
in the search results for the phrase anchor on the day of writing this article,
there is only the subpage page.html, where the word can be found in the
link’s anchor. While Googling the phrase testhash, our domain does not
rank either.
Initially, GB was interested in this funny part of the URL just after the
query mark and the anchor inside the anchor link.
Intrigued, GB was trying to figure out what I meant. It thought, “Is it a
riddle?” To avoid indexing the duplicate content under the other URLs, the
canonical page.html was pointing at itself. e logs altogether registered
crawls on this address, but the conclusions were rather sad:
After weeks, the frequency of GB’s visits decreased significantly until it
eventually left and never crawled that link again.
page.html wasn’t indexed under the phrase anchor, nor was the
parameter with the URL parameter. According to Search Console, this
link does not exist (it is not counted among incoming links), but at the
same time, the phrase anchor is listed as an anchored phrase.
Li k t
https://outline.com/yWqNy6
b it f di ti 8/15
Link to a website from a redirection
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
I wanted to force GB to crawl my website more, which resulted in GB,
every couple of days, entering the dofollow link with an anchor anchor on
page.html leading to page.html, which redirects with a code to
page.html. Unfortunately, as in the case of the page with a parameter, after
days page.html was not yet ranking in the search results for the anchor
phrase which appeared in the redirected link on page.html.
Here’s what happened when I followed Googlebot for 3 months
However, in Google Search Console, in the Anchor Texts section, anchor is
visible and indexed. is could indicate that, after a while, the redirection
will begin to function as expected, so that page.html will rank in the search
results for anchor despite being the second link to the same target page
within the same website.
< link rel=“canonical” href=”https://example.com/page.html” />
is test gave the following results:
. e link for the anchor phrase directing to page.html redirecting
canonically to page.html was not transferred to the target page (just like
in the other cases).
. page.html was indexed despite the canonical tag.
. page.html did not rank in the search results for anchor.
. page.html ranked on the phrases used in the page’s text, which indicated
https://outline.com/yWqNy6 9/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
that GB totally ignored the canonical tags.
I would venture to claim that using rel=canonical to prevent the indexing of
some content (e.g. while filtering) simply could not work.
While designing an SEO strategy, I wanted to make GB dance to my tune
and not the other way around. To this aim, I verified the SEO processes on
the level of the server logs (access logs and error logs) which provided me
with a huge advantage. anks to that, I knew GB’s every movement and
how it reacted to the changes I introduced (website restructuring, turning
the internal linking system upsidedown, the way of displaying information)
within the SEO campaign.
One of my tasks during the SEO campaign was to rebuild a website in a way
that would make GB visit only those URLs that it would be able to index
and that we wanted it to index. In a nutshell: there should only be the pages
that are important to us from the point of view of SEO in Google’s index.
On the other hand, GB should only crawl the websites that we want to be
indexed by Google, which is not obvious to everyone, for example, when an
online shop implements filtering by colors, size and prices, and it is done by
manipulating the URL parameters, eg.:
example.com/women/shoes/?color=red&size=&price=
It may turn out that a solution which allows GB to crawl dynamic URLs
makes it devote time to scour (and possibly index) them instead of crawling
the page.
https://outline.com/yWqNy6 10/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
example.com/women/shoes/
Such dynamically created URLs are not only useless but potentially harmful
to SEO because they can be mistaken for thin content, which will result in
the drop of website rankings.
Within this experiment I also wanted to check some methods of structuring
without using rel=”nofollow”, blocking GB in the robots.txt file or placing
Here’s what happened when I followed Googlebot for 3 months
part of the HTML code in frames that are invisible for the bot (blocked
iframe).
I tested three kinds of JavaScript links.
< a href=”javascript:void()” onclick=”window.location.href =’page.html’”
>anchor< /a >
GB easily moved on to the subpage page.html and indexed the entire page.
e subpage does not rank in the search results for the anchor phrase, and
this phrase cannot be found in the Anchor Texts section in Google Search
Console. e conclusion is that the link did not transfer the juice.
To summarize:
A classic JavaScript link allows Google to crawl the website and index the
pages it comes upon.
It does not transfer juice – it is neutral.
https://outline.com/yWqNy6 11/15
Javascript link with an internal function
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
I decided to raise the game but, to my surprise, GB overcame the obstacle in
less than hours after the publication of the link.
< a href=”javascript:void()” class=”jslink” dataurl=”page.html” >anchor<
/a >
Here’s what happened when I followed Googlebot for 3 months
To operate this link, I used an external function, which was aimed at reading
the URL from the data and the redirection – only the redirection of a user,
as I hoped – to the target page.html. As in the earlier case, page.html had
been fully indexed.
What is interesting is that despite the lack of other incoming links,
page.html was the third most frequently visited page by GB in the entire
web service, right after page.html and page.html.
I had used this method before for structuring web services. However, as we
can see, it does not work anymore. In SEO nothing lives forever, apart from
the Yellow Pages.
< a href=”javascript:void()” class=”jslink” data
url=”cGFnZTEwLmhbWw=” >anchor< /a >
As a result, GB was unable to produce a JavaScript code that would both
https://outline.com/yWqNy6 12/15
30/11/2018
, p J p
Here’s what happened when I followed Googlebot for 3 months
decode the content of a dataURL attribute and redirect. And there it was!
We have a way to structure a web service without using rel=nonfollows to
prevent bots from crawling wherever they like! is way, we do not waste
our crawlbudget, which is especially important in the case of big web
services, and GB finally dances to our tune. Whether the function was
introduced on the same page in the head section or an external JS file, there
is no evidence of a bot either in the server logs or in Search Console.
In the final test, I wanted to check whether the content in, for example,
hidden tabs would be considered and indexed by GB or whether Google
rendered such a page and ignored the hidden text, as some specialists have
been claiming.
I wanted to either confirm or dismiss this claim. To do that, I placed a wall
of text with over signs on page.html and hid a block of text with
about percent of the text ( signs) in Cascading Style Sheets and I
added the show more button. Within the hidden text there was a link to
page.html with an anchor anchor.
ere is no doubt that a bot can render a page. We can observe it in both
Google Search Console and Google Insight Speed. Nevertheless, my tests
revealed that a block of text displayed after clicking the show more button
was fully indexed. e phrases hidden in the text ranked in the search results
and GB was following the links hidden in the text. Moreover, the anchors of
the links from a hidden block of text were visible in Google Search Console
in the Anchor Text section and page.html also began to rank in the search
results for the keyword anchor.
https://outline.com/yWqNy6 lf l h h f l d h dd b 13/15
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
is is crucial for online shops, where content is often placed in hidden tabs.
Now we are sure that GB sees the content in hidden tabs, indexes them, and
transfers the juice from the links that are hidden there.
e most important conclusion that I am drawing from this experiment is
that I have not found a direct way to bypass the First Link Counts Rule by
using modified links (links with parameter, redirects, canonicals, anchor
links). At the same time, it is possible to build a website’s structure using
Here’s what happened when I followed Googlebot for 3 months
Javascript links, thanks to which we are free from the restrictions of the First
Link Counts Rule. Moreover, Google Bot can see and index content hidden
in bookmarks and it follows the links hidden in them.
Opinions expressed in this article are those of the guest author and not necessarily
Search Engine Land. Staff authors are listed here.
We're listening.
Have something to say about this article? Share it with us on Facebook,
Twitter or our LinkedIn Group.
http://outline.com/yWqNy6 COPY
https://outline.com/yWqNy6 14/15
Annotations · Report a problem
30/11/2018 Here’s what happened when I followed Googlebot for 3 months
https://outline.com/yWqNy6 15/15