- I want to cluster big data set (more than 1M records).
- I want to use
dbscan
orhdbscan
algorithms for this clustering task.
When I try to use one of those algorithms, I'm getting memory error.
- Is there a way to fit big data set in parts ? (go with for loop and refit every 1000 records) ?
- If no, is there a better way to cluster big data set, without upgrading the machine memory ?