Deconvolving BOLD activation in event-related designs for multivoxel pattern classification analyses

Jeanette A Mumford, Benjamin O Turner, F Gregory Ashby, Russell A Poldrack, Jeanette A Mumford, Benjamin O Turner, F Gregory Ashby, Russell A Poldrack

Abstract

Use of multivoxel pattern analysis (MVPA) to predict the cognitive state of a subject during task performance has become a popular focus of fMRI studies. The input to these analyses consists of activation patterns corresponding to different tasks or stimulus types. These activation patterns are fairly straightforward to calculate for blocked trials or slow event-related designs, but for rapid event-related designs the evoked BOLD signal for adjacent trials will overlap in time, complicating the identification of signal unique to specific trials. Rapid event-related designs are often preferred because they allow for more stimuli to be presented and subjects tend to be more focused on the task, and thus it would be beneficial to be able to use these types of designs in MVPA analyses. The present work compares 8 different models for estimating trial-by-trial activation patterns for a range of rapid event-related designs varying by interstimulus interval and signal-to-noise ratio. The most effective approach obtains each trial's estimate through a general linear model including a regressor for that trial as well as another regressor for all other trials. Through the analysis of both simulated and real data we have found that this model shows some improvement over the standard approaches for obtaining activation patterns. The resulting trial-by-trial estimates are more representative of the true activation magnitudes, leading to a boost in classification accuracy in fast event-related designs with higher signal-to-noise. This provides the potential for fMRI studies that allow simultaneous optimization of both univariate and MVPA approaches.

Copyright © 2011 Elsevier Inc. All rights reserved.

Figures

Figure 1
Figure 1
The model estimation approaches considered for obtaining trial-by-trial parameter estimates. Five of the approaches (least squares, two versions of ridge regression, partial least squares and support vector regression) used the design matrix shown on the top left, XS. This design matrix contains a single regressor for each trial (in this case 10) in the run, where each regressor is an impulse function convolved with a double gamma HRF. The middle design corresponds to the Add6 model, which models each trial using an unconvolved boxcar function capturing the time point 6 seconds after the time of stimulus presentation. The last design illustrates the LS-S approach, where a trial-specific design matrix is used to obtain the activation estimate for that trial. The design matrices contain two regressors, one for the trial of interest plus a second that models all other trials simultaneously. So, XT1 is the design to obtain the activation estimate for trial 1 and has a regressor modeling that trial and a second regressor modeling all other trials. The estimate for β1 from this first design is the activation for trial 1. This process is repeated N times to obtain estimates for all trials. The bottom table lists the regularization parameters used, when needed.
Figure 2
Figure 2
Illustration of the double cross-validation that was used in the simulation study. The primary cross-validation was a leave-one-run-out CV across 3 runs, for the purposes of obtaining classification accuracy reflecting the ability of the trial-by-trial parameter estimates to predict task type. One fold of the primary cross-validation is illustrated here. It begins with the parameter selection CV, which is a 2-fold CV used to select the regularization parameters used in ridge regression, PLS and SVR. Once this 2-fold CV is carried out the trial-specific activations can be estimated for all 3 runs. Then the primary CV is carried out, training a classifier to predict trial type using 2 runs of data and then testing this model on the test data.
Figure 3
Figure 3
Accuracy results from simulation studies. Each simulation consisted of 500 iterations and 60 trials split randomly and evenly between two tasks. Each of the 12 sets of boxplots contain the results from the eight different models outlined in Figure 1 and arranged by decreasing levels of collinearity (left to right) and increasing noise (top to bottom). When the noise is lower (top 2 rows) the Add6 model performs the worst across all levels of collinearity. For low noise, high collinearity cases (upper left) the LS-S model outperforms all other models. The solid horizontal line indicates chance (50%), while the dashed line indicates accuracy significantly better than chance according to the binomial distribution.
Figure 4
Figure 4
Correlation of trial-by-trial parameter estimates with true parameter estimate value from simulation studies. The correlations between true and estimated trial activation magnitude were similar for β1 and β2 and hence are averaged for each simulation. The organization of the boxplots is similar to Figure 3 with collinearity decreasing from left to right across rows and noise increasing from top to bottom. Methods with weaker correlation between estimated activation and true activation magnitudes tended to have worse classification results.
Figure 5
Figure 5
Variability of the parameter estimates. For each simulation the variance of the parameter estimates across trials was calculated and variances were similar for β1 and β2, so they are averaged for each simulation. As expected the unregulated LS-A model has the highest variance, although in some cases the SVR parameter estimates are much more variable. The variability of LS-S is often similar to the other regulated approaches, but is found to be slightly larger than some of the regulated approaches in higher noise situations.
Figure 6
Figure 6
Classification accuracies from the real data analysis. The Add6 accuracy was significantly lower than LS-S (p=0.0002, uncorrected) and Add4–6 (p

Figure 7

Classification accuracies from the real…

Figure 7

Classification accuracies from the real data analysis separated by stimulus type. Note that…

Figure 7
Classification accuracies from the real data analysis separated by stimulus type. Note that the classification accuracy increase for LS-S and Add4–6 compared to the other methods, as shown in Figure 6, is mostly due to an increase in the classification accuracy of the mirrored words.
All figures (7)
Figure 7
Figure 7
Classification accuracies from the real data analysis separated by stimulus type. Note that the classification accuracy increase for LS-S and Add4–6 compared to the other methods, as shown in Figure 6, is mostly due to an increase in the classification accuracy of the mirrored words.

References

    1. Carlson TA, Schrater P, He S. Patterns of activity in the categorical representations of objects. J Cogn Neurosci. 2003;15:704–717.
    1. Chen X, Pereira F, Lee W, Strother S, Mitchell T. Exploring predictive and reproducible modeling with the single-subject FIAC dataset. Hum Brain Mapp. 2006;27:452–461.
    1. Cox DD, Savoy RL. Functional magnetic resonance imaging (fMRI) ”brain reading”: detecting and classifying distributed patterns of fMRI activity in human visual cortex. Neuroimage. 2003;19:261–270.
    1. De Martino F, Valente G, Staeren N, Ashburner J, Goebel R, Formisano E. Combining multivariate voxel selection and support vector machines for mapping and classification of fMRI spatial patterns. Neuroimage. 2008;43:44–58.
    1. Hastie T, Tibshirani R, Friedman J. The elements of statistical learning. Springer; 2001.
    1. Haxby JV, Gobbini MI, Furey ML, Ishai A, Schouten JL, Pietrini P. Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science. 2001;293:2425–2430.
    1. Hoerl A, Kennard R. Ridge regression: applications to nonorthogonal problems. Technometrics. 1970;12:69–82.
    1. Kamitani Y, Tong F. Decoding seen and attended motion directions from activity in the human visual cortex. Curr. Biol. 2006;16:1096–1102.
    1. Kriegeskorte N, Mur M, Bandettini P. Representational similarity analysis - connecting the branches of systems neuroscience. Front Syst Neurosci. 2008;2:4.
    1. LaConte S, Strother S, Cherkassky V, Anderson J, Hu X. Support vector machines for temporal classification of block design fMRI data. Neuroimage. 2005;26:317–329.
    1. Misaki M, Kim Y, Bandettini PA, Kriegeskorte N. Comparison of multivariate classifiers and response normalizations for pattern-information fMRI. Neuroimage. 2010;53:103–118.
    1. Mourao-Miranda J, Reynaud E, McGlone F, Calvert G, Brammer M. The impact of temporal compression and space selection on SVM analysis of single-subject and multi-subject fMRI data. Neuroimage. 2006;33:1055–1065.
    1. Mur M, Bandettini PA, Kriegeskorte N. Revealing representational content with pattern-information fMRI–an introductory guide. Soc Cogn Affect Neurosci. 2009;4:101–109.
    1. Pereira F, Mitchell T, Botvinick M. Machine learning classifiers and fMRI: a tutorial overview. Neuroimage. 2009;45:199–209.
    1. Polyn SM, Natu VS, Cohen JD, Norman KA. Category-specific cortical activity precedes retrieval during memory search. Science. 2005;310:1963–1966.
    1. Rissman J, Gazzaley A, D’Esposito M. Measuring functional connectivity during distinct stages of a cognitive task. Neuroimage. 2004;23:752–763.
    1. Scholkopf B, Smola A. Learning with kernels. MIT press; 2002.
    1. Turner B. A comparison of methods for the use of pattern classification on rapid event-related fMRI data. Poster session presented at the Annual Meeting of the Society for Neuroscience; San Diego, CA. 2010.
    1. Woolrich MW, Ripley BD, Brady M, Smith SM. Temporal autocorrelation in univariate linear modeling of FMRI data. Neuroimage. 2001;14:1370–1386.
    1. Xue G, Dong Q, Chen C, Lu Z, Mumford JA, Poldrack RA. Greater neural pattern similarity across repetitions is associated with better memory. Science. 2010;330:97–101.

Source: PubMed

3
Abonnieren