Chung-Ki Wong1, Vadim Zotev1, Raquel Phillips1, Masaya Misaki1, and Jerzy Bodurka1,2,3
1Laureate Institute for Brain Research, Tulsa, OK, United States, 2Stephenson School of Biomedical Engineering, University of Oklahoma, Norman, OK, United States, 3Center for Biomedical Engineering, University of Oklahoma, Norman, OK, United States
Synopsis
In
EEG-fMRI, EEG electrodes record head motions with a high temporal resolution
(EEG-motion-sensor), which can be utilized for retrospective slice-by-slice fMRI
motion correction. EEG motion components derived from independent component (IC)
analysis were automatically identified by the common features observed in the
IC mean power spectral density, spatial projection topographic map, and signal
contribution. For real-time application of the EEG-motion-sensor,
pre-trained models
are desirable for faster classification.
We used convolutional neural network to evaluate performance of motion-IC classification
model. High speed and classification accuracy were achieved on a large EEG-fMRI
dataset, suggesting the possibility of real-time EEG-motion-sensor applications
for fMRI.
Introduction
In
simultaneous EEG-fMRI, ability of EEG cap electrodes to record and measure head
motion artifacts at a high temporal resolution (EEG-motion-sensor) can be
utilized to substantially remove head movements on a slice-by-slice basis in
the fMRI data.1,2 EEG motion components derived from independent
component analysis (ICA)3 were automatically identified by the common
features observed in the mean power spectral density, spatial projection
topographic map, and signal contribution of the independent components (ICs).1
For
potential real-time application of the EEG-motion-sensor for fMRI
slice-by-slice head motion corrections, pre-trained models are desirable for faster classification.
Here we used convolutional neural network (ConvNet)4 to evaluate the
performance of motion-IC classification model. In ConvNet, the convolutional
layers filter local patterns of input features regardless of their positions, which
improve feature recognition efficiency. Unlike the identifications based on
threshold parameters, a machine learning approach benefits from a large IC
dataset without the need for parameters adjustment. It also allows convenient model
updates by updating the ICs database with new data.Methods
The
simultaneous EEG-fMRI were conducted on a GE MR750 3T MRI with an 8ch head coil
and 32ch MR-compatible EEG (Brain Products GmbH). The study included 1040 scans
from 56 subjects. Each scan lasted for 526s with the first 6s removed to ensure
the fMRI signal steady state. For EEG preprocessing, MRI artifacts were
removed. ICA was performed on the preprocessed data to separate 31 ICs for each
scan. A total of 32240 ICs were obtained. For motion-ICs classification, 92
features were used. They were 31 IC spatial projection components (A), 50 points in the normalized mean
power spectral density below 12Hz, and 11 IC statistical properties:5
max(|IC(t)|), max(IC(t))-min(IC(t)), max(|IC'(t)|), stdev(IC(t)), skew(IC(t)),
kurt(IC(t)), mean(IC(t)2)*mean(A2), mean(var(IC(τ1))), mean(var(IC(τ2))), mean(skew(IC(τ1))), and mean(skew(IC(τ2))).6 Here t spans the scan duration, τ1=1s and τ2=10s are durations of moving periods, and IC' is
the IC time derivative. It is noted that simultaneous changes of the signs of A and IC do not affect the EEG signal
(Fig.1). To increase the number of datasets for better classification, the
number of datasets is doubled by flipping the signs of A and IC (i.e., the skewness calculations of the ICs), giving a
total of 64480 datasets: 38688 for training, 12896 for cross-validation, and
12896 for testing. The train and cross-validation datasets were used for model
training and optimization, and the test dataset was separated for model evaluation.
In the ConvNet, two convolutional layers followed by a hidden layer and a logistic
regression layer were used. The cost function was L2-regularized. The classification
results were evaluated on the test dataset by accuracy (Acc), precision (P), recall (R), and F1-score.
They were compared with those calculated by support vector machine (SVM) method,6,7,8
and 1- or 2-hidden-layer multilayer perception neural network (ML1/ML2),9
which were essentially the ConvNet with the convolution layers removed (ML1) or
replaced by another hidden layer (ML2). The neural network calculations were carried
out using Theano in Python.10Results
It
takes about 4.5s to classify the test dataset (0.35ms/IC on average) on
Intel® Xeon® workstation at 2.50GHz using 10 threads.
Figs.2-4 plot (Acc,P,R,F1) against the learning rate (LR) for
different regularization parameters (L2) for ConvNet, ML1 and ML2. In the plots,
(Acc,P,R,F1) achieve (0.976,0.910,0.888,0.899), (0.972,0.912,0.854,0.882),
and (0.972,0.904,0.863,0.883) for ConvNet, ML1 and ML2 respectively. The precisions of ML1 and ML2 are comparable to that
of ConvNet, but the recall, and hence F1,
are significantly lower. There are two types of head motion-ICs: the spontaneous
and cardioballistic motion-ICs.1,2 When the two types of motion-ICs
were classified by two separate SVM models, (Acc,P,R,F1)2-models=(0.967,0.876,0.851,0.863). When one
model was used, (Acc,P,R,F1)1-model=(0.970,0.882,0.874,0.878). Among all
models, ConvNet gives the highest accuracy for the motion-ICs classification.Discussions and Summary
Motion-ICs
from EEG-fMRI were classified based on a database of 64480 ICs (7612
motion-related) without presetting or adjusting threshold parameters using the
convolutional neural network. The average time to classify one IC dataset is
about 0.35ms. The result achieves high speed and classification accuracy (97.6%), which suggests the
possibility of the EEG-motion-sensor application for real-time slice-by-slice
fMRI head motion corrections.Acknowledgements
This
work is supported by DOD award W81XWH-12-1-0607.References
1. Wong CK,
Zotev V, Misaki M, et al. An automatic EEG-assisted retrospective motion correction
for fMRI (aE-REMCOR). NeuroImage. 2016;129:133-147.
2. Zotev V, Yuan
H, Phillips R., et al. EEG-assisted retrospective motion correction for fMRI:
E-REMCOR. NeuroImage. 2012;63:698-712.
3. Bell AJ and
Sejnowski TJ. An information-maximization approach to blind separation and
blind deconvolution. Neural Comput. 1995;7:1129-1159.
4. LeCun Y,
Bottou Y, Bengio Y, et al. Gradient-based learning applied to document
recognition. Proc. IEEE 1998;86:2278-2324.
5. Winkler I.
Automatic classification of artifactual ICA-components for artifact removal in
EEG signals. Behavioral and Brain Functions 2011;7:30.
6. Wong CK,
Zotev V, Misaki M, et al. Support vector machine classification of head motion
independent components from EEG-fMRI. Proc. Org. Hum. Brain Map 2016;22:1811.
7. Cortes C. and
Vapnik V. Support-vector networks. Machine Learning 1995;20:273-297.
8. Christianini
N. and Shawe-Taylor J.C. An
introduction to support vector machines and other kernel-based learning methods.
Cambridge University Press, Cambridge, UK. 2000.
9. Christopher M
Bishop. Pattern recognition and machine learning. 2006; Section 5.
10. Theano
Development Team. Theano: A Python framework for fast computation of
mathematical expressions. May 2016. arXiv:1605.02688 [cs.SC].