I have a small magento site which consists of page URLs such as:
http://www.example.com/contact-us.html
http://www.example.com/customer/account/login/
However I also have pages which include filters (e.g. price and colour) and two such examples are:
http://www.example.com/products.html?price=1%2C1000
http://www.example.com/products/chairs.html?price=1%2C1000
The issue is that when Google bot and the other search engine bots search the site, it essentially grinds to a halt because they get stuck in all the "filter links".
So, in the robots.txt
file how can it be configured e.g:
User-agent: *
Allow:
Disallow:
To allow all pages like:
http://www.example.com/contact-us.html
http://www.example.com/customer/account/login/
to get indexed but in the case of http://www.example.com/products/chairs.html?price=1%2C1000
index products.html
, but ignore all the content after the ?
?
The same should apply for http://www.example.com/products/chairs.html?price=1%2C1000
I also don't want to have to specify each page, in turn just a rule to ignore everything after the ?
but not the main page itself.