Dear HCP experts,
I attempted the structural connectome tutorial using the individual HCP
dataset (100307) with MRtrix. I am looking to repeat a similar
'whole-brain' tractography analysis using FSL probtrackx. Specifically, I
would like to utilize the freesurfer aparc+aseg file as nodes for
What pipelines did you run on this data?
Peace,
Matt.
From:
>
on behalf of Joey Contreras
>
Date: Tuesday, October 24, 2017 at 5:45 PM
To:
Hello
I have acquired resting state AP, PA data and have ran it through functional
HCP pipeline. I am looking to load and analyze in connectome workbench but
cannot find the .dconn or deserves.nii files. Would I have had to run another
script to generate these files? Specifically I had hoped
For a quick introduction to CIFTI, see FAQ #1:
https://wiki.humanconnectome.org/display/PublicData/HCP+Users+FAQ
You may also want to read wb_command -cifti-help, while some of it is
redundant with the FAQ, it also explains things you need to know about when
using cifti commands in wb_command:
GIFTI is really only for single structures. CIFTI is for combinations across
structures. Perhaps you should create a CIFTI file with wb_command
-cifti-create-dense-timeseries.
Peace,
Matt.
From:
>
on
Hi,
I'd like to merge two gifti files, each of format 32K x t timepoints.
I'd like the output to be 64K x t timepoints, so basically concatenating
the first dimension.
My understanding is that wb_command -metric-merge doesn't acomodate this.
Is there anything else I can try?
I tried converting to
Hi,
As I was exploring your data, especially the Source-MEG time series, I noticed
the time-series’ length is 14960 pts for each vertex. However, at the
channel-level analysis, the time series are of 140 000 pts, with fs of 1034 Hz.
I would like to know why the number of points at the source
Hi Meizhen,
Participants completed the Language task eyes open as in all the HCP tfMRI.
Greg, can you speak to any important considerations or effects?
Best,
Jenn
Jennifer Elam, Ph.D.
Scientific Outreach, Human Connectome Project
Washington University School of Medicine
Department of
Thanks very much!
-- 原始邮件 --
发件人: "Harms, Michael" ;
发送时间: 2017年10月24日(星期二) 22:35
收件人: "罗" <963619...@qq.com>;
抄送: "hcp-users" ;
主题: Re: [HCP-Users] about the hcp data
Yes, all 1113 subjects with MR data
Hi David,
Thanks for the reply and the explanation, glad to hear it is coming
along and sorry for the technical issues.
Bryan.
On 24/10/2017 13:58, David Van Essen wrote:
> Bryan et al.
>
> Apologies for the delay in releasing the reprocessed 7T data. In
> brief, it’s still not done, and it
Yes, all 1113 subjects with MR data in the S1200 release are available to use.
There are a small number of issues that we discovered in the S500 (and S900)
release data after the initial release, which were fixed as part of those
subjects’ data for the S1200 release.
See here for the details:
I'm sorry.
I question is that whether all the 1113 subjects with MRI are available for our
research, because some s500 subjects are not included in our brought S500
connectome box.
Thanks.
-- 原始邮件 --
发件人: "Harms, Michael" ;
发送时间:
Hi,
Can you rephrase your question? Are you asking if the S1200 data release is
available for purchase as part of the “Connectome in a Box” program?
If so, the answer is yes:
https://store.humanconnectome.org/data/
cheers,
-MH
--
Michael Harms, Ph.D.
13 matches
Mail list logo