Page 11 - Human Environment Interface (4)
P. 11
Brain Reading for Human-Machine Interfaces
ML Analysis. All ML evaluations have been performed using and 600–900 ms) and subject. Corrections were applied where
the open source signal processing and classification environment necessary. Classification performance after optimizing the classifier
pySPACE [66]. Data processing was as follows: Windowing and were analyzed using repeated measures ANOVA with subject as
preprocessing were performed directly on the raw data from the within-subjects factor. Where necessary, the Greenhouse–Geisser
recording device. In order to avoid that preprocessing artifacts correction was applied and the corrected p-value is reported. For
such as, e.g., filter border artifacts, influence classification multiple comparisons, the Bonferroni correction was applied.
performance, we performed the complete preprocessing (including
decimation and filtering) on a larger window between {200 and In a further analysis we investigated the possibility to improve
1400 ms relative to the stimulus onset. We chose the following classification performance by the combination of information from
preprocessing based on the rationale issued above (see Sec. two windows. We combined the middle time window (300 to
‘‘Average ERP Analysis’’ and [64]): The data were baseline- 600 ms) with both other time windows (early: 0 to 300 ms and
corrected (with 100 ms window prior to stimulus onset), decimated late: 600 to 900 ms time window) separately.
to 25 Hz and subsequently lowpass filtered with a cut-off
frequency of 4 Hz. To determine classification performance that can be achieved
under optimized conditions we finally performed a final analysis
As in the ERP analysis, run numbers 2, 3 and 4 of both sessions with the goal to get a better estimate of the applicability of our
were used for training and testing. In contrast to the average ERP approach of classifier transfer between the classes standard and
analysis described above we included the early time window of 0– missed target. The processing window was chosen based on the
300 ms in the ML analysis. This was done to control for the fact results of the systematic ML analysis explained above. We
that early time windows may still contribute to the classification of performed a classifier optimization of the SVM parameter
the different classes (standards, targets, missed targets) even though we complexity using a 5-fold cross validation in combination with a
hypothesized that main differences are caused by the P300 effect pattern search algorithm [69] to evaluate the overall performance
(see Sec. ‘‘ Relevant Averaged ERP Activity’’). in the application with an adjusted classifier.
It is important to note here that several dependencies have to be In this subsequent investigation we used an optimized SVM on
kept in mind when evaluating the results: First, performance a chosen time window. Further, we computed the balanced
depends on window size since a larger window contains more accuracy (BA) as a performance measure for the chosen time
features and thus a higher dimensionality of the signal. Second, the window. The balanced accuracy [70] is the arithmetic mean of
classifier parameters depend on the underlying data (and true positive rate (TPR) and true negative rate (TNR) and
dimensionality). However, the purpose of this investigation was calculated accordingly
to compare different windows (and sizes) concerning their quality
for classification, so we assessed the results always with respect to BA~ 1 ðTPRzTNRÞ: ð1Þ
window size and starting point of the window and performed 2
statistical analysis only on windows of equal length.
Both performance measures used (AUC and BA) are insensitive
Furthermore, the parameters of the classifier were adjusted to to unbalanced or even changing ratios of the two classes (positive
an unspecified value to omit data-dependent effects: in the entire class P and negative class N), which is most important in the
analysis, we used a support vector machine (SVM) as implemented application where we have an oddball-like situation with frequent
in LIBSVM [67] (SVC-C with a linear kernel) with a fixed and infrequent examples. It holds for both metrics that a value of
complexity of 100 simulating a hard margin. Hence, we cut 0:5 means guessing and 1 means perfect classification.
different windows by varying starting point (0 ms-700 ms) and
window size (200 ms-800 ms) in steps of 100 ms. Data used for Results
training and testing were different, as outlined above: We trained Behavior. In total 724 omission errors (575:2+82:23)
on standards and targets of one experimental run and tested missed
targets versus targets of another run within one session. All possible occurred, thus 724 missed targets were observed and 2876 targets
combinations of the above mentioned runs within one session were stimuli were found with correct responses. No commission error
tested. Classifier features were the preprocessed time-channel (i.e., responses on standards stimuli) could be found.
values, i.e., the amplitudes.
Figure 7 shows the median response time for each subject across
The corresponding classification performance was computed two sessions. Based on the buzzer press event, responses occurred
using the area under curve (AUC) [68] which is an indicator of 837 ms after the target stimuli (mean of subject’s medians). The
general separability of the two classes in the data. AUC is the area median of minimal response time was 597 ms and the median of
under the receiver operating characteristics curve. This curve maximal response time was 1783 ms. The difference between the
maps the different true positive rates (TPR) and false positive rates minimal and maximal response time was between 686 ms and
(1-TNR) obtained when the decision boundary is varied from 1657 ms (median: 1131 ms). EMG onsets began even earlier in
{? to ?. The AUC is then computed as the integral of the time (mean of subject’s medians: 551 ms). The median of minimal
resulting function. In this way, we investigated the linear response time was 336 ms and the median of maximal response
separability of the data essentially independent of the applied time was 1466:5 ms. The difference between the minimal and
classifier. maximal response time was between 563 ms and 1621 ms
(median: 1130:5 ms). No difference exists between median
For statistical inference, we chose three time windows from the difference in response time based on the buzzer event (median:
aforementioned temporal segmentation that match the later time 1131 ms) and median difference of response time based on the
windows which had been chosen for ERP analysis (300–600 ms, EMG onset (median: 1130:5 ms).
and 600–900 ms see Sec. ‘‘Average ERP Analysis’’) and the early
time window of 0–300 ms. This procedure relates the results of the Relevant Averaged ERP Activity. The grand average over
classifier performance-based approach to the results of the ERP all subjects of the standard, target and missed target ERP pattern
analysis. Classification performances for the different window sizes in the centro-parietal electrode (Pz) is depicted in Fig. 4.
were statistically analyzed using repeated measures ANOVA with Significant differences between standards and targets (i.e., P300
the within-subjects factors time window (0–300 ms, 300–600 ms, effect) were observed [F (2,50)~65:27,pv0:001, pairwise com-
PLOS ONE | www.plosone.org 9 December 2013 | Volume 8 | Issue 12 | e81732

