On Sun, Apr 15, 2012 at 9:43 PM, Christopher Brown <[email protected]> wrote:
> Hi, > > My apologies if members don't consider this list appropriate, if so > perhaps a more appropriate forum could be suggested. > > I am using scientific linux 6.2. I will be installing it on 4-5 computers > that will make up my lab. I plan to add some additional packages from > source or otherwise: > > /opt/python2.7 > /opt/python3.2 > /opt/libreoffice3.5 > /opt/zotero3.0 > /opt/qt-4.8.1 > > ...and a few more plus a number of python2.7 packages such as numpy, pyqt, > etc. I have my environment setup pretty well on one machine with a bash > script I wrote to download make and install all the packages I need. It > also sets up a decent /etc/skel/, so that global settings are set for all > users. I would like all the machines to be identical, and I would like to > be able to update packages automatically. My question is, what is the best > way to deploy and maintain this environment onto several networked > machines? Setting up my own repo seems like a lot of overhead for a few > machines. Is there a better way? I could go around to each machine with my > bash script, but what would the best way be to handle updates? I am willing > to learn about repository management if necessary, but obviously the less > effort/knowledge required the better, as I am a relative novice at system > administration. Thus, any suggestions will be appreciated. > you need a configuration management tool. The 'cool' kids appear to choose puppet, we use cfengine (like facebook or amd, to name just two huge deployments). Whatever you choose, running a config management tool will be an ongoing project on your part, it requires a change of mindset on the admin's part. Some can, some cannot. -- groet, Natxo
