The Welsh Advanced Neuroimaging Database (WAND)
Carolyn Beth McNabb1, Ian D Driver1, Vanessa Hyde1, Garin Hughes1, Hannah Louise Chandler1, Hannah Thomas1, Eirini Messaritaki1, Carl Hodgetts1,2, Craig Hedge3, Christopher Allen4, Maria Engel1, Sophie Felicity Standen1, Emma L Morgan1, Elena Stylianopoulou1, Svetla Manolova1, Lucie Reed1, Mark Drakesmith1, Michael Germuska1, Alexander D Shaw5, Lars Mueller6, Holly Rossiter1, Christopher W Davies-Jenkins7,8, Tom Lancaster 9, John Evans1, David Owen1, Gavin Perry1, Slawomir Kusmia1,10, Emily Lambe1, Adam M Partridge1, Allison Cooper1, Peter Hobden1, Hanzhang Lu7,8, Kim S Graham1,12, Andrew D Lawrence1,12, Richard G Wise1,13,14, James T R Walters15, Petroc Sumner1, Krish D Singh1, and Derek K Jones1
1Cardiff University Brain Research Imaging Centre, School of Psychology, Cardiff University, Cardiff, United Kingdom,
2Department of Psychology, Royal Holloway, University of London, Egham, United Kingdom,
3School of Psychology, Aston University, Birmingham, United Kingdom,
4Department of Psychology, Durham University, Durham, United Kingdom,
5Washington Singer Laboratories, University of Exeter, Exeter, United Kingdom,
6Leeds Institute of Cardiovascular and Metabolic Medicine, University of Leeds, Leeds, United Kingdom,
7The Russel H. Morgan Department of Radiology and Radiological Science, Johns Hopkins University School of Medicine, Baltimore, Maryland, USA,
8F.M. Kirby Research Center for Functional Brain Imaging, Kennedy Kreiger Institute, Baltimore, Maryland, USA,
9Department of Psychology, University of Bath, Bath, United Kingdom,
10IBM Polska Sp. z o. o., Department of Content Design, Cracow, Poland
11University of Sheffield, Research Services, New Spring House, 231 Glossop Road, Sheffield, S10 2GW, UK
12School of Philosophy, Psychology and Language Sciences, Dugald Stewart Building, University of Edinburgh, 3 Charles Street, Edinburgh, EH8 9AD, UK
13Department of Neurosciences, Imaging, and Clinical Sciences, 'G. D'Annunzio' University of Chieti-Pescara, Chieti, Italy,
14Institute for Advanced Biomedical Technologies, 'G. D'Annunzio' University of Chieti-Pescara, Chieti, Italy
15School of Medicine, Centre for Neuropsychiatric Genetics and Genomics, Cardiff University, Cardiff, United Kingdom
Acknowledgements:
The WAND data were acquired at the UK National Facility for In Vivo MR Imaging of Human Tissue Microstructure funded by the EPSRC (grant EP/M029778/1), and The Wolfson Foundation, and supported by a Wellcome Trust Investigator Award (096646/Z/11/Z) and a Wellcome Trust Strategic Award (104943/Z/14/Z). The UK Medical Research Council (MR/M008932/1), the Welsh Government and Cardiff University provide funding support for the CUBRIC ultra-high field imaging facility.
The authors would like to thank:
Thomos Wizel (formally MGH, now GBIO) for his help setting up the diffusion sequences on the Connectom scanner.
Olivier Mougin (University of Nottingham) for developing the phase-sensitive inversion recovery and T1 code.
William Clarke (Oxford University) for providing the image reconstruction code for the MP2RAGE.
Many researchers in the Cardiff University Brain Research Imaging Centre (CUBRIC) contributed their time and expertise throughout the project. The authors would like to especially thank Mara Cercignani and Marco Palombo for their help with quantitative and diffusion imaging questions, the CUBRIC IT team for their help with storage, and data organisation, and the CUBRIC reception team for their help with participant admin.
Dataset description:
The Welsh Advanced Neuroimaging Database (WAND) is a multi-scale, multi-modal imaging database of the healthy human brain. It includes non-invasive in vivo brain data from 3T MRI with ultra-strong (300mT/m) magnetic field gradients (especially designed for evaluating tissue microstructure), 7T and 3T MRI and nuclear magnetic resonance spectroscopy (MRS), and MEG. The dataset also includes cognitive and behavioural data that allow for investigation of brain-behaviour interactions at multiple spatial and temporal scales.
Data naming structure and organisation:
Data have been organised using the Brain Imaging Data Structure (BIDS). Where BIDS specifications were not available, we have consulted with experts in the field to align as closely as possible with this data structure.
BIDS naming conventions for many of the Connectom (ses-02) acquisitions are still in their infancy - we have used names as close as we can to the existing rules and bent these slightly where recommended by experts.
qMT
FlipAngle and MTOffsetFrequency for quantitative MRI are labelled according to specifications outlined in the qMRLab BIDS recommendations. See here. For qMT data, we have used the mt-
entity to index the MTOffsetFrequency
- this is against the official BIDS specification (on/off only) but allowed for better definition between qMT scans.
N.B. mt-1
will refer to the first MTOffsetFrequency for that flip angle (e.g. flip-1
), meaning that flip-1_mt-1
could refer to a different MTOffsetFrequency than flip-2_mt-1
. For details on flip angle and MT offset frequency, refer to the .json file for that image.
mcDESPOT
mcDESPOT images are labelled according to specifications outlined in the Quantitative MRI BIDS documentation here. For mcDESPOT data, we have used the acq-
entity to index between spgr
, spgr_IR
and SSFP
acquisitions. We have used the Variable Flip Angle VFA
modality label to define mcDESPOT scans.
Thanks to Agah Karakuzu for recommendations on BIDS formatting for quantitiative MRI.
AxCaliber
For AxCaliber and CHARMED, we have used the acq-
entity to index the scan sequence: AxCaliber
or CHARMED
. Both have dwi
as their modality label.
Big delta values for the AxCaliber scans are added to the .json file as t_bdel
in keeping with recommendations from the aDWI-BIDS preprint. The acq-
label is used to define which AxCaliber scan the image came from. acq-AxCaliber1
refers to the "AX_delta17p3_30xb2200_30xb4400" scan, acq-AxCaliber2
to the "AX_delta30_30xb4000_30xb8000" scan and so on. The P>A reference scan is referred to as acq-AxCaliberRef
.
Some data will fail the BIDS validation
This is relevant for the MEG, MRS and TMS data, where data are stored as .mat or .dat files. However, the folder structure and naming convention will still be in keeping with the rest of the dataset.
Data availability:
Raw data have been made available, along with brain extracted T1 images, and quality metrics for multishell diffusion, T1, T2 and functional (BOLD) data. These are available in the derivatives
folder of our GitLab repository..
Here we provide instructions for downloading and interacting with the WAND data using GIN. GIN will allow you to interrogate and work with the data without requiring terabytes of storage space.
If you are having trouble accessing the dataset, please send an email to cubricwand@cardiff.ac.uk
External user instructions
The WAND data are hosted on the G-Node GIN repository. To access the data, please follow these instructions.
Checklist (have you done all these steps?)
- Sign up to the G-Node GIN service using the Register button at the top of the Homepage (or click here)
- Make sure you have
conda
installed. See here for installation instructions.
- Install
git-annex
(using conda
) using these instructions
- Install the
gin
command-line client following these instructions
- Add your public SSH key here and sign in with the client using
gin login
.
You will now be able to clone the repository to your machine. First, change into the directory where you want the repository to download. In the terminal window, type:
cd </path/to/newhome>
And then use gin to pull down the repository:
gin get CUBRIC/WAND
cd WAND
cat README.md
By default, large files will not be downloaded automatically. You will be able to see them but not open them. To download larger files, use the command:
gin get-content <large-file>
And to remove downloaded files:
gin remove-content . # remove all downloaded files
gin remove-content folder/ # remove a particular folder
You should now have access to the full WAND dataset.
PLEASE BE AWARE THAT THE DATASET IS HUGE - WE DO NOT ADVISE USING gin get-content *
UNLESS YOU HAVE A LOT OF SPACE AND A LOT OF TIME TO WAIT FOR THE DATA TO FINISH DOWNLOADING. Instead, we recommend only downloading the data you need, as you need it.
Additional scripts
In addition to the data release itself, we have shared some code, used to run quality control on the diffusion data using FSL's eddy_qc and the T1, T2 and BOLD functional data using MRIqc. These can be found in the code
folder of our GitLab repository.
Physiological data
Physiological monitoring data for some functional scans are currently unavailable. The research team is working hard to ensure these data are made available as soon as possible.