I am working with the Python runtime (if that matters for this). I have been struggling with the number_found_accuracy. The documentation on the api is a bit lacking for this particular field. The name of it seems to suggest one would set it to a number like 200 for instance, meaning "try to be accurate to within 200". However, reading other's accounts and given that this value has a max value (in the api MAXIMUM_NUMBER_FOUND_ACCURACY) of 10,000 currently, this suggests that you set the value for a number over what you expect to get and the search mechanism does it's best with accuracy.
My problem is that I am doing a search that I expect to return around 32,000 results. I set the number_found_accuracy to the max, but the results have a number_found that varies and seems to be way off. For example, returning 90,000. I am using a cursor by the way. Is there a way to find out how many total documents exist in an index independently of number_found? More generally, how does one debug situations like this? And finally, is the number_found property of search results useful if there are over 10k or so results? If not, what are others using on GAE, integrating and calling out to BigQuery?
Thanks for any insight.