Hi again, 

So, we just found out that one of our automatically launched scripts had a bug 
and was creating library datasets over and over again. Ending up in 16000+ 
entries of the same library dataset before it started to fail. 
That's not at all a solution to the problem but brings me to another question 
:) 

Because we cannot use the web interface anymore to delete them, is there a way 
to remove all these library datasets from the database? 
Would it work to flag them deleted and use the purge scripts to locate and 
remove the associated files? And after that, which database tables do we have 
to clean in order to remove the library datasets completely from the database? 


Thanks a lot, 

Sajoscha

On Jan 29, 2013, at 3:47 PM, Sajoscha Sauer wrote:

> Dear all, 
> 
> We have a really annoying problem with one of our Libraries: we can't access 
> it anymore. We think the problem might be that the library contains too many 
> folders/files (the error message is pasted at the end of the mail). 
> 
> Is there any limit for library content ? Note that we can still browse 
> smaller libraries, so I though that we should simply split the library into 
> smaller ones but since we can t even browse it...Any suggestions how to 
> approach this problem?
> 
> We are running a galaxy instance used by different research groups and our 
> current setup is the following: 
> - we have one Library per research group
> - we create a folder per project, under the relevant library, and add all 
> files in these folders. Note that we add symbolic links to the real file 
> locations (this way the large fastq files remain on owners file servers)
> - the files are transferred automatically to Galaxy using custom code that 
> makes use of the galaxy API. 
> 
> And of course, this automated transfer now crashes as our custom code tries 
> to read library content thru API call (to ask the user in which folder the 
> new files should be added)
> 
> PS: browsing libraries in general is really slow and gives users a bad 'user 
> experience' (not mentioning all the extra clicks they issue because they 
> think Galaxy did not get the request...), I am not sure why this action is 
> that slow ( browsing small size library is also slow). I think I saw emails 
> about this issue but I am not sure what the status is?  Is there anything we 
> could do in the code to speed this up? I am not sure this problem is related 
> to the current post , so sorry for mixing things up.
> 
> Thanks for your help!
> 
> Cheers
> 
> Sajoscha
> 
> 
> 
> === error on galaxy page (nothing in the galaxy log file)
> 
> <error.tiff>
> 
> ___________________________________________________________
> Please keep all replies on the list by using "reply all"
> in your mail client.  To manage your subscriptions to this
> and other Galaxy lists, please use the interface at:
> 
>  http://lists.bx.psu.edu/

___________________________________________________________
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Reply via email to