On Oct 24, 2011, at 2:57 AM, Peter Cock wrote:

> On Mon, Oct 24, 2011 at 4:53 AM, John Chilton <chil...@msi.umn.edu> wrote:
>> Hello All,
>>     This is more of a general policy question. The head of our Research
>> Informatics Support Systems group here at the University of Minnesota wants
>> to know how other institutions are handling the large amount of data
>> generated by next generation sequencing facilities and processed by local
>> Galaxy instances. Specifically, I think she is interested in what sort of
>> data retention policies institutions have in place - for instance how long
>> do you store data, what sort of quotas do you put in place, what sort of
>> backups do you have in place, etc....
>> Any comments would be appreciated.
>> Thanks for your time,
>> -John
> Might be worth asking on the bioinformatics core facilities mailing list
> too, I know some of the folk there use Galaxy in their institutes:
> http://open-bio.org/mailman/listinfo/bioinfo-core
> Peter

We're in the (very long, drawn-out) process of integrating much of our local 
bioinformatics resources to also deal with this problem.  We have a general 
plan in place, including adding automated tape backup and a number of other 
features on the analysis end (storage for fast I/O, etc), but I'm interested in 
hearing more from others on this.


Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:


Reply via email to