I have access to a group of servers connected with a 1Gb LAN, and each of them has 40+ cores and Ubuntu OS. They all have a common NAS. I installed SLURM on a few of them and configured it so that each server is both a control and a compute node, and the servers are not connected. The required analyses are bioinformatic and are CPU bound but with files in GB. My questions are as following:
- This is not a compute cluster, correct? What would be needed to link these servers into a single cluster?
- Is it a valid practice to use HPC managers like Slurm on this configuration? How would the data be shared? They do have a common NAS, but running any computations on the NAS directly is very slow compared to the local files.
My ideal solution would pull the files to a local machine (ideally regardless of their location, but NAS could be the common hub) perform the computation and possibly return the output files. Is this an unreasonable request or a solved problem?
Thanks in advance!