Thanks for the reply. I would like to reduce the computing time for DiffusionPreprocessing; feeding in both sets of dMRIs takes about 24.16 hours to run, while parallel processing takes 10.72 hours. I've been trying to retrieve the eddy_cuda version, but I can't find it in the neurodebian packages. Where can I find pre-compiled binaries of eddy_cuda? The platform I am using in my Docker is Ubuntu 14.04 LTS.
Thank you again, Yeun On Mon, Jul 17, 2017 at 11:57 AM, Harms, Michael <mha...@wustl.edu> wrote: > > Hi, > Is there a particular reason that you can’t provide all the dMRI scans at > once, and let the pipeline handle the merging for you? > If you process each dMRI run separately, then the individual runs will not > be in optimal alignment. (You would be relying on the registration of each > run to the T1, rather than registering the dMRI directly to each other as > part of ‘eddy’). > > cheers, > -MH > > -- > Michael Harms, Ph.D. > ----------------------------------------------------------- > Conte Center for the Neuroscience of Mental Disorders > Washington University School of Medicine > Department of Psychiatry, Box 8134 > 660 South Euclid Ave. Tel: 314-747-6173 <(314)%20747-6173> > St. Louis, MO 63110 Email: mha...@wustl.edu > > From: <hcp-users-boun...@humanconnectome.org> on behalf of Yeun Kim < > yeun...@gmail.com> > Date: Monday, July 17, 2017 at 1:32 PM > To: "Glasser, Matthew" <glass...@wustl.edu> > > Cc: "hcp-users@humanconnectome.org" <hcp-users@humanconnectome.org> > Subject: Re: [HCP-Users] diffusion data merge pipeline > > I am using the following function (and is looped through the pairs of > unique sets of gradient tables (i.e. loops twice for dir99 and dir98): > ${HCPPIPEDIR}/DiffusionPreprocessing/DiffPreprocPipeline.sh \ > --posData="{posData}" \ > --negData="{negData}" \ > --path="{path}" \ > --subject="{subject}" \ > --echospacing="{echospacing}" \ > --PEdir={PEdir} \ > --gdcoeffs="NONE" \ > --dwiname="{dwiname}" \ > --printcom=""' > > Where: > $posData = diffusion data in the positive direction > $negData = diffusion data in the negative direction > $path = output directory path > $echospacing = echospacing > $PEdir = 2 > $dwiname = i.e. Diffusion_dir-98_run-01 > > > FYI: I'm using HCPPipelines v3.17. > > ------------------------------------------------------------- > > Technical details: > > I run ${HCPPIPEDIR}/DiffusionPreprocessing/DiffPreprocPipeline.sh in a > Docker container with the following python code. It is looped through the > pairs of unique sets of gradient tables (i.e. loops twice for dir99 and > dir98) and set to process in parallel: > > dwi_stage_dict = OrderedDict([("DiffusionPreprocessing", > partial(run_diffusion_processsing, > > *posData*=pos, > > *negData*=neg, > > *path*=args.output_dir, > > *subject*="sub-%s" % subject_label, > > *echospacing*=echospacing, > > *PEdir*=PEdir, > > *gdcoeffs*="NONE", > > *dwiname*=dwiname, > > *n_cpus*=args.n_cpus))]) > for stage, stage_func in > dwi_stage_dict.iteritems(): > if stage in args.stages: > Process(target=stage_func).start() > > On Mon, Jul 17, 2017 at 11:15 AM, Glasser, Matthew <glass...@wustl.edu> > wrote: > >> The pipeline is capable of doing the merge for you if you want. Can you >> post how you called the diffusion pipeline? >> >> Peace, >> >> Matt. >> >> From: Yeun Kim <yeun...@gmail.com> >> Date: Monday, July 17, 2017 at 1:12 PM >> To: Matt Glasser <glass...@wustl.edu> >> Cc: "hcp-users@humanconnectome.org" <hcp-users@humanconnectome.org> >> Subject: Re: [HCP-Users] diffusion data merge pipeline >> >> When I run DiffusionPreprocessing, I make the --dwiname=DWIName specific >> to the diffusion scan (i.e. DWIName= Diffusion_dir-98_run-01) to prevent >> files from being overwritten. >> I end up with: >> ${StudyFolder}/${Subject}/T1w/Diffusion_dir-98_run-01/data.nii.gz >> ${StudyFolder}/${Subject}/T1w/Diffusion_dir-99_run-01/data.nii.gz >> >> I would like to combine the two data.nii.gz files. >> >> On Mon, Jul 17, 2017 at 10:58 AM, Glasser, Matthew <glass...@wustl.edu> >> wrote: >> >>> Look for the ${StudyFolder}/${Subject}/T1w/Diffusion/data.nii.gz file. >>> >>> Peace, >>> >>> Matt. >>> >>> From: <hcp-users-boun...@humanconnectome.org> on behalf of Yeun Kim < >>> yeun...@gmail.com> >>> Date: Monday, July 17, 2017 at 12:56 PM >>> To: "hcp-users@humanconnectome.org" <hcp-users@humanconnectome.org> >>> Subject: [HCP-Users] diffusion data merge pipeline >>> >>> Hi, >>> >>> We have the diffusion scans: >>> dMRI_dir98_AP, dMRI_dir98_PA >>> dMRI_dir99_AP, dMRI_dir99_PA >>> >>> in which there is a pair of phase encoding directions (AP,PA) and two >>> sets of different diffusion weighting directions (dir98 and dir99). >>> >>> After running the DiffusionPreprocessing module of the HCP minimal >>> preprocessing pipeline, I would like to merge the processed dMRI_dir98 and >>> dMRI_dir99 data. Do you have any suggestions on how to perform this step? >>> Also, are there any workflows developed by HCP for >>> post-DiffusionPreprocessing? >>> >>> Thank you, >>> Yeun >>> >>> _______________________________________________ >>> HCP-Users mailing list >>> HCP-Users@humanconnectome.org >>> http://lists.humanconnectome.org/mailman/listinfo/hcp-users >>> >> >> > _______________________________________________ > HCP-Users mailing list > HCP-Users@humanconnectome.org > http://lists.humanconnectome.org/mailman/listinfo/hcp-users > > > ------------------------------ > > The materials in this message are private and may contain Protected > Healthcare Information or other information of a sensitive nature. If you > are not the intended recipient, be advised that any unauthorized use, > disclosure, copying or the taking of any action in reliance on the contents > of this information is strictly prohibited. If you have received this email > in error, please immediately notify the sender via telephone or return mail. > _______________________________________________ HCP-Users mailing list HCP-Users@humanconnectome.org http://lists.humanconnectome.org/mailman/listinfo/hcp-users