Awesome, I’ll look into that all. What do you mean by exports some scratch space?
Thanks, Trevor > On May 7, 2015, at 11:04 PM, Uwe Sauter <[email protected]> wrote: > > > Trevor, > > if you have the possibility: set up an extra machine that > > a) manages users via LDAP > b) exports user homes via NFS > c) exports some scratch space (though that won't scale performance-wise) > > and you ship around both the topics you asked on this list. > > > Regards, > > Uwe > > > Am 07.05.2015 um 20:27 schrieb Trevor Gale: >> >> Thank you for your detailed response. I think my main issue is that I’m very >> new to Slurm, and clusters in general. I plan on setting up a global file >> system across my desktops, and was wondering what software you would >> recommend. I saw that the Slurm documentation mentions Lustre and NFS but >> was just curious because I have no experience with either. >> >> Thanks, >> Trevor >> >>> On May 7, 2015, at 7:28 PM, Uwe Sauter <[email protected]> wrote: >>> >>> >>> Trevor, >>> >>> I don't know what your intent is or the machine you are preparing yourself >>> for but in general login nodes and compute nodes share >>> a common filesystem, making the need to move data around (inside of the >>> cluster) unnecessary. >>> >>> If you really need to move data from node local space back to the login >>> node, there are several possibilities to do so: >>> >>> * Export some part of the login node's filesystem to your compute node. >>> * Put a SCP/RSYNC into your job script. (Make sure you're SSH keys are >>> placed in the authorized_keys file) >>> * Run a "data mover" job that depends on your compute job (and the node >>> where the compute job ran). >>> >>> Likely there are more solutions to your problem. But before you go any >>> further it'd be good if you put some thought into your >>> setup. Does it represent what you are trying to achieve? >>> >>> Regards, >>> >>> Uwe >>> >>> >>> >>> Am 07.05.2015 um 15:19 schrieb Trevor Gale: >>>> >>>> Hello, >>>> >>>> I’m currently running one desktop computer as a controller and one as a >>>> compute node for testing. I’m running a simple test script using salloc >>>> and then passing the script over sbcast to my node where i execute it >>>> calling srun. The problem I’m having is that I want the output of the >>>> programs I’m going to run to come back to the head node (or dump there in >>>> the first place) after execution, but all out my outputs are dropping onto >>>> the node that they execute on. Does slurm support any method of output >>>> collection? or is there some configuration I can change to move all the >>>> outputs to the head node? This seems like an issue that other users would >>>> encounter, does anyone have a good method for fixing this? >>>> >>>> Thanks, >>>> Trevor >>>>
