It sounds like a simple question. Exclude the waybackmachine crawler (ia_archiver
) and allow all other user agents.
So I setup the robots.txt as follows:
User-agent: *
Sitemap: https://www.example.com/sitemap.xml
User-agent: ia_archiver
Disallow: /
After half a year I noticed that the visitor count to my site dropped tremendously.
After a while I realized that Google Bot stopped indexing my site.
Confirmed by their robots.txt verifier:
The Disallow: /
part is picked up by google bot too, not only ia_archiver
is blocked.
The obvious question is:
What is wrong with this robots.txt?
Is the order of the entries the culprit?