Hi Chris, sounds good to me.
On 26.06.2012, at 17:03, Christopher Brooks <[email protected]> wrote: > Tobias, > > I'd like to amend this to be even more clear about the performance: > > # The directory where the system stores files and directories that > # should be shared with other Matterhorn nodes. This is usually > # a networked file location. This location should provide hard-linking > # functionality, and provide large amounts of disk space. See [ref] > # for an estimate. > # > # In the case of a single machine install, the setting may be left as > # is. > org.opencastproject.storage.shared.dir=${java.io.tmpdir}/shared > > # The directory where the system stores files that are unique to > # this particular node within a Matterhorn cluster. This directory > # will store temporary files as well as persistent ones, so chose a > # place that is persistent between reboots. > # > # > org.opencastproject.storage.local.dir=${java.io.tmpdir}/${org.opencastproject.server.hostname} > > The most important part of this is indicating which one needs to store > a lot of space versus a little space. And the hard linking issue. > Ideally we would include a link to the wiki right inline in the > configuration file, this would have helped out our systems group a lot. > > Chris > >> 1) Introduce two distinct variables for local and shared storage. >> >> # The directory where the system stores files and directories that >> should be shared with other Matterhorn # nodes within the same >> cluster for performance and space saving reasons. # >> # Note that by default, the directory points to each node's >> ${java.io.tmpdir}/shared directory, which should # be changed to a >> mounted filesystem that is shared between multiple Matterhorn nodes. >> In the case of a # one box install, the setting may be left as is. >> org.opencastproject.storage.shared.dir=${java.io.tmpdir}/shared >> >> # The directory where the system stores files that are unique to >> every node within a Matterhorn cluster. # >> # Note that this directory will store temporary files as well as >> persistent ones, so make sure to chose # a place that is persistent >> between reboots. >> org.opencastproject.storage.local.dir=${java.io.tmpdir}/${org.opencastproject.server.hostname} >> >> 2) Use these two variables in the appropriate places. >> >> The new variables will be used for default values of potentially >> shared places (workspace, working file repository) as well as >> decidedly local (read: per node) directories such as the search >> indexes. >> >> e. g >> org.opencastproject.workspace.rootdir=${org.opencastproject.storage.shared.dir}/workspace >> org.opencastproject.workingfiles.rootdir=${org.opencastproject.storage.shared.dir}/files >> >> and >> org.opencastproject.search.solr.dir=${org.opencastproject.storage.local.dir}/searchindex >> ... >> >> 3) Remove the existing (and no longer used) variable >> org.opencastproject.storage.dir >> >> This variable needs to be removed both from config.properties itself >> as well as from some parts of the code. This certainly sounds like >> work, but having adopter installs that screw up their installation >> because they are overwriting search indexes because the went to the >> same shared storage location is not helpful for the project. >> >> Tobias >> _______________________________________________ >> Matterhorn mailing list >> [email protected] >> http://lists.opencastproject.org/mailman/listinfo/matterhorn >> >> >> To unsubscribe please email >> [email protected] >> _______________________________________________ > > > > -- > Christopher Brooks, BSc, MSc > ARIES Laboratory, University of Saskatchewan > > Web: http://www.cs.usask.ca/~cab938 > Phone: 1.306.966.1442 > Mail: Advanced Research in Intelligent Educational Systems Laboratory > Department of Computer Science > University of Saskatchewan > 176 Thorvaldson Building > 110 Science Place > Saskatoon, SK > S7N 5C9 _______________________________________________ Matterhorn mailing list [email protected] http://lists.opencastproject.org/mailman/listinfo/matterhorn To unsubscribe please email [email protected] _______________________________________________
