I have uploaded around 1TB of data on elasticsearch DB. For searching I tried following ways -
"from+size" that has default value of index.max_result_window as 10000, but I wanted to search from 100000, hence I set index.max_result_window to 100000. Then searched from 100000 and size=10, but it causes heap size full.
Scroll API - For keeping older segments alive utilizes more file handles. hence it again consumes the memory configured in nodes.
search_after - I tried sorting documents on basis of _uid, but it gives me follwoing error -
-
{
"error": {
"root_cause": [
{
"type": "circuit_breaking_exception",
"reason": "[fielddata] Data too large, data for [_uid] would be [13960098635/13gb], which is larger than the limit of [12027297792/11.2gb]",
"bytes_wanted": 13960098635,
"bytes_limit": 12027297792
}
}
},
What can be done to resolve this error and also which is the most efficient way to search a large chunk of data through pagination?