I'm looking for a robots.txt parser in Java, which supports the same pattern matching rules as the Googlebot.
I've found some librairies to parse robots.txt files, but none of them supports Googlebot-style pattern matching :
- Heritrix (there is an open issue on this subject)
- Crawler4j (looks like the same implementation as Heritrix)
- jrobotx
Does anyone know of a java library that can do this ?