On Oct 24, 2011, at 2:57 AM, Peter Cock wrote: > On Mon, Oct 24, 2011 at 4:53 AM, John Chilton <chil...@msi.umn.edu> wrote: >> Hello All, >> >> This is more of a general policy question. The head of our Research >> Informatics Support Systems group here at the University of Minnesota wants >> to know how other institutions are handling the large amount of data >> generated by next generation sequencing facilities and processed by local >> Galaxy instances. Specifically, I think she is interested in what sort of >> data retention policies institutions have in place - for instance how long >> do you store data, what sort of quotas do you put in place, what sort of >> backups do you have in place, etc.... >> >> Any comments would be appreciated. >> >> Thanks for your time, >> -John > > Might be worth asking on the bioinformatics core facilities mailing list > too, I know some of the folk there use Galaxy in their institutes: > http://open-bio.org/mailman/listinfo/bioinfo-core > > Peter
We're in the (very long, drawn-out) process of integrating much of our local bioinformatics resources to also deal with this problem. We have a general plan in place, including adding automated tape backup and a number of other features on the analysis end (storage for fast I/O, etc), but I'm interested in hearing more from others on this. chris ___________________________________________________________ Please keep all replies on the list by using "reply all" in your mail client. To manage your subscriptions to this and other Galaxy lists, please use the interface at: http://lists.bx.psu.edu/