GithubHelp home page GithubHelp logo

datalad-datasets / hcp_movies Goto Github PK

View Code? Open in Web Editor NEW
0.0 3.0 1.0 24.36 MB

This subset comprises subjects with movie data from the WU-Minn HCP1200 dataset, including the relevant 7T movie, 7T rest, and 3T T1w data. More info in README.

Shell 100.00%

hcp_movies's Introduction

Get data from a subset (7T movie data) from the Human Connectome Project Open Access dataset with DataLad

made-with-datalad

This dataset enables data retrieval for a subset of the HCP Open Access dataset for users that accepted the WU-Minn HCP Consortium Open Access Data Use Terms and obtained valid AWS credentials via db.humanconnectome.org.

Movie data subset

Important: For a DataLad dataset of the full HCP Open Access dataset, please go to github.com/datalad-datasets/human-conntectome-project-openaccess.

This subset comprises the 7T movie, 7T rest, and 3T T1w_MPR1 data files for each participant with 7T data in the HCP Open Access dataset. Specifically, these files are

- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/filescans.csv
- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/<sub>_7T_tfMRI_MOVIE*_{AP,PA}_SBRef.nii.gz
- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/<sub>_7T_tfMRI_MOVIE*_{AP,PA}.nii.gz
- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/<sub>_7T_SpinEchoFieldMap_{AP,PA}.nii.gz
- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/LINKED_DATA/BEHAV/<sub>_7T_MOV*_behav.xml
- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/LINKED_DATA/EYETRACKER/<sub>_7T_MOV*_eyetrack.asc
- <sub>/unprocessed/7T/tfMRI_MOVIE*_{AP,PA}/LINKED_DATA/EYETRACKER/<sub>_7T_MOV*_eyetrack_summary.csv
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/filescans.csv
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/<sub>_7T_rfMRI_REST*_{AP,PA}_SBRef.nii.gz
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/<sub>_7T_rfMRI_REST*_{AP,PA}.nii.gz
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/<sub>_7T_SpinEchoFieldMap_{AP,PA}.nii.gz
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/LINKED_DATA/BEHAV/<sub>_7T_REST*_behav.xml
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/LINKED_DATA/EYETRACKER/<sub>_7T_REST*_eyetrack.asc
- <sub>/unprocessed/7T/rfMRI_REST*_{AP,PA}/LINKED_DATA/EYETRACKER/<sub>_7T_REST*_eyetrack_summary.csv
- <sub>/unprocessed/3T/T1w_MPR1/<sub>_3T_T1w_MPR1.nii.gz
- <sub>/unprocessed/3T/T1w_MPR1/<sub>_3T_FieldMap_Phase.nii.gz
- <sub>/unprocessed/3T/T1w_MPR1/<sub>_3T_FieldMap_Magnitude.nii.gz
- <sub>/unprocessed/3T/T1w_MPR1/<sub>_3T_BIAS_BC.nii.gz
- <sub>/unprocessed/3T/T1w_MPR1/<sub>_3T_BIAS_32CH.nii.gz
- <sub>/unprocessed/3T/T1w_MPR1/<sub>_3T_AFI.nii.gz

The directory structure and the file names in this subset are kept identical to the full HCP dataset. An additional BIDS-formatted view of the dataset is available in the bids branch.

The purpose of this dataset is to give easy access to a single dataset with the 7T movie data.

Human Connectome Project

The Human Connectome Project (HCP) aims to construct a map of the complete structural and functional neural connections in vivo within and across individuals.

Its 'WU-Minn HCP Open Access Data' data release includes high-resolution 3T MR scans from young healthy adult twins and non-twin siblings (ages 22-35) using four imaging modalities: structural images (T1w and T2w), resting-state fMRI (rfMRI), task-fMRI (tfMRI), and high angular resolution diffusion imaging (dMRI). It further includes behavioral and other individual subject measure data for all, and MEG data and 7T MR data for a subset of subjects (twin pairs).

Data access and retrieval with DataLad

To retrieve HCP Open Access data via DataLad with this dataset, you need to agree to the WU-Minn HCP Consortium Open Access Data Use Terms and obtain valid AWS credentials:

  • Create an account at http://db.humanconnectome.org.
  • Log into your account and accept the data use terms of the "WU-Minn HCP Data - 1200 Subjects" data release.
  • Enable Amazon S3 access for your Amazon account to get an access key ID and a secret access secret key (click on the button with the S3 logo).
    • The access key ID is a character string similar to this: AKIAXOX5CT57GHZ4SVFV
    • The secret access key is a character string similar to this: vntFcVA+YI0Ii3tVZPpdyQrgp2H05YjesyXKGE+n

You will be asked to supply your AWS credentials the first time you use datalad get to retrieve file content of your choice from the HCP Open Access dataset. You should only need to provide credentials once, and all subsequent datalad get commands will retrieve data without asking them again.

Dataset structure

Each HCP1200/ subject directory in this dataset is a DataLad subdataset. The command datalad get -n <subject-id> clones this subdataset and allows to access this subjects release notes (subdirectory release-notes). Within each subject's subdataset, one DataLad subdataset exists for each additional available subdirectory (e.g., MEG, T1w, etc., as far as available for the particular subject).

If you have never used DataLad before, please read the section on DataLad datasets below.

DataLad datasets and how to use them

This repository is a DataLad dataset. It allows fine-grained data access up to the level of single files of the HCP Open Access dataset without hosting the HCP data. In order to use this repository for data retrieval, DataLad is required. It is a free and open source command line tool, available for all major operating systems, and builds up on Git and git-annex to allow sharing, synchronizing, and version controlling collections of large files. You can find information on how to install DataLad at handbook.datalad.org/en/latest/intro/installation.html.

Get the dataset

A DataLad dataset can be cloned by running

datalad clone <url>

Once a dataset is cloned, it is a light-weight directory on your local machine. At this point, it contains only small metadata and information on the identity of the files in the dataset, but not actual content of the (sometimes large) data files.

Retrieve dataset content

After cloning a dataset, you can retrieve file contents by running

datalad get <path/to/directory/or/file>

This command will trigger a download of the files, directories, or subdatasets you have specified.

DataLad datasets can contain other datasets, so called subdatasets. If you clone the top-level dataset, subdatasets do not yet contain metadata and information on the identity of files, but appear to be empty directories. In order to retrieve file availability metadata in subdatasets, run

datalad get -n <path/to/subdataset>

Afterwards, you can browse the retrieved metadata to find out about subdataset contents, and retrieve individual files with datalad get. If you use datalad get <path/to/subdataset>, all contents of the subdataset will be downloaded at once.

Stay up-to-date

DataLad datasets can be updated. The command datalad update will fetch updates and store them on a different branch (by default remotes/origin/master). Running

datalad update --merge

will pull available updates and integrate them in one go.

More information

More information on DataLad and how to use it can be found in the DataLad Handbook at handbook.datalad.org. The chapter "DataLad datasets" can help you to familiarize yourself with the concept of a dataset.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.