This is really a bioinformatics question, but I'll make it as general as I can. Here's the semi-hypothetical situation:
Let's say I have access to a cluster or even a cloud. I want to run some very specific programs on this cluster/cloud (genomic/transcriptomic assembly programs to be exact). The problem is that I expect these programs (Velvet/Oases, Trinity, whatever else) to require large amounts of RAM, conservatively 100GB+, and the largest node on my cluster/cloud is only 32GB.
Now besides switching to an MPI/Hadoop based program (ABySS or whatnot), writing my own, or buying a new computer, what are my viable options? Has anyone tried using a distributed operating system (MOSIX, Kerrighed, ...) with shared memory on multiple nodes of a cluster/cloud? What about a virtual SMP? What else?
Thanks for any help!
Edit for clarification: Let's also say that the programs mentioned above (Velvet/Oases and Trinity) require a single system with a large pool of RAM. In a nut shell, I'm looking for a viable way to "paste" a bunch of nodes together into one virtual super-node where a single process could access all of the RAM from all of the nodes like it was a single system. I know that anything like this would probably give a pretty substantial performance hit, but I'm looking for something that's possible, not necessarily efficient.
p.s. Sorry if my terminology is making things confusing. I'm somewhat new to a lot of this.