In a proposed project the plan is to divide large data sets into chunks, expose them by SFTP into a swarm CPU type of LAN cluster architecture for processing of historical numerical and financial data for simulations. Time consuming.
What about using something like Influxdb? I know about sharing and scaling a bit. Could we take advantage of this in how it distributes data to nodes and perform processing on them?