- Anatomical data preprocessing
-
A total of 1 T1-weighted (T1w) images were found within the input
BIDS dataset. A preprocessed T1w image was provided as a precomputed
input and used as T1w-reference throughout the workflow. A pre-computed
brain mask was provided as input and used throughout the workflow.
Precomputed discrete tissue segmentations were provided as inputs.
Precomputed tissue probabiilty maps were provided as inputs.
- Preprocessing of B0 inhomogeneity mappings
-
A total of 1 fieldmaps were found available within the input BIDS
structure for this particular subject. A deformation field to correct
for susceptibility distortions was estimated based on SDCFlows’
fieldmap-less approach. The deformation field is that resulting
from co-registering the EPI reference to the same-subject’s
T1w-reference (Wang et al. 2017; Huntenburg
2014). Registration is performed with
antsRegistration (ANTs 2.6.2), and the process regularized
by constraining deformation to be nonzero only along the phase-encoding
direction.
- Functional data preprocessing
-
For each of the 1 BOLD runs found per subject (across all tasks and
sessions), the following preprocessing was performed. First, a reference
volume was generated from the shortest echo of the BOLD run, using a
custom methodology of fMRIPrep, for use in head motion
correction. Head-motion parameters with respect to the BOLD reference
(transformation matrices, and six corresponding rotation and translation
parameters) are estimated before any spatiotemporal filtering using
mcflirt (FSL
, Jenkinson et al. 2002). The estimated fieldmap
was then aligned with rigid-registration to the target EPI (echo-planar
imaging) reference run. The field coefficients were mapped on to the
reference EPI using the transform. The BOLD reference was then
co-registered to the T1w reference using mri_coreg
(FreeSurfer) followed by flirt (FSL , Jenkinson and Smith 2001) with the
boundary-based registration (Greve and Fischl 2009) cost-function.
Co-registration was configured with six degrees of freedom. A
T2★ map was estimated from the preprocessed EPI echoes using
tedana’s t2smap workflow (DuPre et al. 2021), by voxel-wise fitting
the maximal number of echoes with reliable signal in that voxel to a
monoexponential signal decay model with nonlinear regression. The
T2★/S0 estimates from a log-linear regression fit
were used for initial values. The calculated T2★ map was then
used to optimally combine preprocessed BOLD across echoes following the
method described in (Posse
et al. 1999). The optimally combined time series was carried
forward as the preprocessed BOLD. Several confounding
time-series were calculated based on the preprocessed BOLD:
framewise displacement (FD), DVARS and three region-wise global signals.
FD was computed using two formulations following Power (absolute sum of
relative motions, Power et al. (2014)) and Jenkinson
(relative root mean square displacement between affines, Jenkinson et al. (2002)).
FD and DVARS are calculated for each functional run, both using their
implementations in Nipype (following the definitions by Power et al.
2014). The three global signals are extracted within the CSF, the
WM, and the whole-brain masks. Additionally, a set of physiological
regressors were extracted to allow for component-based noise correction
(CompCor, Behzadi
et al. 2007). Principal components are estimated after high-pass
filtering the preprocessed BOLD time-series (using a discrete
cosine filter with 128s cut-off) for the two CompCor variants:
temporal (tCompCor) and anatomical (aCompCor). tCompCor components are
then calculated from the top 2% variable voxels within the brain mask.
For aCompCor, three probabilistic masks (CSF, WM and combined CSF+WM)
are generated in anatomical space. The implementation differs from that
of Behzadi et al. in that instead of eroding the masks by 2 pixels on
BOLD space, a mask of pixels that likely contain a volume fraction of GM
is subtracted from the aCompCor masks. This mask is obtained by
thresholding the corresponding partial volume map at 0.05, and it
ensures components are not extracted from voxels containing a minimal
fraction of GM. Finally, these masks are resampled into BOLD space and
binarized by thresholding at 0.99 (as in the original implementation).
Components are also calculated separately within the WM and CSF masks.
For each CompCor decomposition, the k components with the
largest singular values are retained, such that the retained components’
time series are sufficient to explain 50 percent of variance across the
nuisance mask (CSF, WM, combined, or temporal). The remaining components
are dropped from consideration. The head-motion estimates calculated in
the correction step were also placed within the corresponding confounds
file. The confound time series derived from head motion estimates and
global signals were expanded with the inclusion of temporal derivatives
and quadratic terms for each (Satterthwaite et al.
2013). Frames that exceeded a threshold of 0.5 mm FD or 1.5
standardized DVARS were annotated as motion outliers. Additional
nuisance timeseries are calculated by means of principal components
analysis of the signal found within a thin band (crown) of
voxels around the edge of the brain, as proposed by (Patriat, Reynolds,
and Birn 2017). All resamplings can be performed with a
single interpolation step by composing all the pertinent
transformations (i.e. head-motion transform matrices, susceptibility
distortion correction when available, and co-registrations to anatomical
and output spaces). Gridded (volumetric) resamplings were performed
using nitransforms, configured with cubic B-spline
interpolation.
The above boilerplate text was automatically generated by fMRIPrep
with the express intention that users should copy and paste this text
into their manuscripts unchanged. It is released under the CC0
license.
Abraham, Alexandre, Fabian Pedregosa, Michael Eickenberg, Philippe
Gervais, Andreas Mueller, Jean Kossaifi, Alexandre Gramfort, Bertrand
Thirion, and Gael Varoquaux. 2014.
“Machine Learning for
Neuroimaging with Scikit-Learn.” Frontiers in
Neuroinformatics 8.
https://doi.org/10.3389/fninf.2014.00014.
Behzadi, Yashar, Khaled Restom, Joy Liau, and Thomas T. Liu. 2007.
“A Component Based Noise Correction Method (CompCor)
for BOLD and Perfusion Based fMRI.”
NeuroImage 37 (1): 90–101.
https://doi.org/10.1016/j.neuroimage.2007.04.042.
DuPre, Elizabeth, Taylor Salo, Zaki Ahmed, Peter A. Bandettini,
Katherine L. Bottenhorn, César Caballero-Gaudes, Logan T. Dowdle, et al.
2021.
“TE-Dependent Analysis of Multi-Echo fMRI with
*Tedana*.” Journal of Open Source Software 6 (66): 3669.
https://doi.org/10.21105/joss.03669.
Esteban, Oscar, Ross Blair, Christopher J. Markiewicz, Shoshana L.
Berleant, Craig Moodie, Feilong Ma, Ayse Ilkay Isik, et al. 2018.
“fMRIPrep 25.3.0.dev12+g35a0b0fc3.” Software.
https://doi.org/10.5281/zenodo.852659.
Esteban, Oscar, Christopher Markiewicz, Ross W Blair, Craig Moodie, Ayse
Ilkay Isik, Asier Erramuzpe Aliaga, James Kent, et al. 2019.
“fMRIPrep: A Robust Preprocessing
Pipeline for Functional MRI.” Nature
Methods 16: 111–16.
https://doi.org/10.1038/s41592-018-0235-4.
Gorgolewski, K., C. D. Burns, C. Madison, D. Clark, Y. O. Halchenko, M.
L. Waskom, and S. Ghosh. 2011.
“Nipype: A Flexible, Lightweight
and Extensible Neuroimaging Data Processing Framework in Python.”
Frontiers in Neuroinformatics 5: 13.
https://doi.org/10.3389/fninf.2011.00013.
Gorgolewski, Krzysztof J., Oscar Esteban, Christopher J. Markiewicz,
Erik Ziegler, David Gage Ellis, Michael Philipp Notter, Dorota Jarecka,
et al. 2018.
“Nipype.” Software.
https://doi.org/10.5281/zenodo.596855.
Greve, Douglas N, and Bruce Fischl. 2009.
“Accurate and Robust
Brain Image Alignment Using Boundary-Based Registration.”
NeuroImage 48 (1): 63–72.
https://doi.org/10.1016/j.neuroimage.2009.06.060.
Huntenburg, Julia M. 2014.
“Evaluating Nonlinear Coregistration of
BOLD EPI and T1w Images.” Master's
Thesis, Berlin: Freie Universität.
https://hdl.handle.net/11858/00-001M-0000-002B-1CB5-A.
Jenkinson, Mark, Peter Bannister, Michael Brady, and Stephen Smith.
2002.
“Improved Optimization for the Robust and Accurate Linear
Registration and Motion Correction of Brain Images.”
NeuroImage 17 (2): 825–41.
https://doi.org/10.1006/nimg.2002.1132.
Jenkinson, Mark, and Stephen Smith. 2001.
“A Global Optimisation
Method for Robust Affine Registration of Brain Images.”
Medical Image Analysis 5 (2): 143–56.
https://doi.org/10.1016/S1361-8415(01)00036-6.
Patriat, Rémi, Richard C. Reynolds, and Rasmus M. Birn. 2017.
“An
Improved Model of Motion-Related Signal Changes in fMRI.” NeuroImage 144, Part A
(January): 74–82.
https://doi.org/10.1016/j.neuroimage.2016.08.051.
Posse, Stefan, Stefan Wiese, Daniel Gembris, Klaus Mathiak, Christoph
Kessler, Maria-Liisa Grosse-Ruyken, Barbara Elghahwagi, Todd Richards,
Stephen R. Dager, and Valerij G. Kiselev. 1999.
“Enhancement of
BOLD-Contrast Sensitivity by Single-Shot Multi-Echo
Functional MR Imaging.” Magnetic Resonance in
Medicine 42 (1): 87–97.
https://doi.org/10.1002/(SICI)1522-2594(199907)42:1<87::AID-MRM13>3.0.CO;2-O.
Power, Jonathan D., Anish Mitra, Timothy O. Laumann, Abraham Z. Snyder,
Bradley L. Schlaggar, and Steven E. Petersen. 2014.
“Methods to
Detect, Characterize, and Remove Motion Artifact in Resting State
fMRI.” NeuroImage 84 (Supplement C): 320–41.
https://doi.org/10.1016/j.neuroimage.2013.08.048.
Satterthwaite, Theodore D., Mark A. Elliott, Raphael T. Gerraty, Kosha
Ruparel, James Loughead, Monica E. Calkins, Simon B. Eickhoff, et al.
2013.
“An improved framework for confound
regression and filtering for control of motion artifact in the
preprocessing of resting-state functional connectivity
data.” NeuroImage 64 (1): 240–56.
https://doi.org/10.1016/j.neuroimage.2012.08.052.
Wang, Sijia, Daniel J. Peterson, J. C. Gatenby, Wenbin Li, Thomas J.
Grabowski, and Tara M. Madhyastha. 2017.
“Evaluation of Field Map
and Nonlinear Registration Methods for Correction of Susceptibility
Artifacts in Diffusion MRI.” Frontiers in
Neuroinformatics 11.
https://doi.org/10.3389/fninf.2017.00017.