On Feb 29, 2012, at 11:34 AM, Jorrit Boekel wrote:
> Dear list,
> Our lab's proteomics data is frequently outputted into >50 files containing
> different fractions of proteins. The files are locally stored and not present
> on the Galaxy server. We've planned to somehow (inside galaxy) merge these
> files and split them into tasks so they can be run on the cluster. We would
> either merge/split the files by concatenation, or untar/tar files at every
> job, depending on filetype and tool. No problems so far.
> However, I have been looking around for a way to upload >50 files
> simultaneously to galaxy and convert to one dataset, and this does not seem
> to be supported. Before starting to create a hack to make this work, which
> doesn't seem especially trivial to me, I'd like to know if I should instead
> use libraries. From what I've seen, libraries are not treated as datasets in
> Galaxy but rather contain datasets. If there was a "tar all sets in library
> and import to history" I'd be using that, but I've only encountered "tar/zip
> sets and download locally" which would be a bit of a workaround.
It's not possible to do this all in one step, but you can definitely upload
them all simultaneously and then concatenate them using the concatenate tool
(or write a simple tool to tar them).
> I haven't found much on this subject in the mailing list, has this
> functionality been requested before?
> jorrit boekel
> Please keep all replies on the list by using "reply all"
> in your mail client. To manage your subscriptions to this
> and other Galaxy lists, please use the interface at:
Please keep all replies on the list by using "reply all"
in your mail client. To manage your subscriptions to this
and other Galaxy lists, please use the interface at: