How to set up a bunch of linux servers with shared file system without using job scheduler?
2
votes
1
answer
317
views
I am managing multiple GPU servers in our lab, which are mainly used for deep learning tasks. We would like these machines to share the same file system, so it is easier to switch between them.
Currently, I am using NFS to share the
/home
folders for all the machines but installing system updates on all of them (like NVIDIA driver, outside of the home folder) is quite painful since I have to do it on each machine.
I wonder if there is any way to share the entire file system (the root /
). I am concerned that since these machines are of different configurations (different CPU, GPU, memory) and running different jobs, and there are folders like /dev
, /proc
, /tmp
. is it a good idea to directly share the /
?
I read some posts on how to set up a Linux cluster and most of them suggest using a scheduling system like Slurm. However, our servers are mainly for algorithm development (debugging) so directly connecting them from client machines is preferred. Is it possible to share the entire filesystem without using a job queue system?
Asked by x.y.z liu
(33 rep)
Apr 13, 2023, 09:18 PM
Last activity: Apr 13, 2023, 11:23 PM
Last activity: Apr 13, 2023, 11:23 PM