I am working on a cluster where I submit jobs through the qsub
engine.
I am granted a maximum of 72h of computational time at once. The output of my simulation is a folder which typically contains about 1000 files (about 10 Gb). I copy my output back after 71h30m of simulation. This means that everything that is produced after 71h30m (+ time to copy?) is lost. Is there a way to make the process more efficient, that is not having to manually estimate the time needed to copy output back?
Also before copying back my output I compress files with bzip2
, what resources are used to do that? Should I ask a 1 node more than what I need to run the simulation only to compress files?