Figures
Abstract
In brain-computer interfaces (BCIs) based on steady-state visual evoked potentials (SSVEPs) the number of selectable targets is rather limited when each target has its own stimulation frequency. One way to remedy this is by combining frequency- with phase encoding. We introduce a new multivariate spatiotemporal filter, based on Linearly Constrained Minimum Variance (LCMV) beamforming, for discriminating between frequency-phase encoded targets more accurately, even when using short signal lengths than with (extended) Canonical Correlation Analysis (CCA), which is traditionally posited for this stimulation paradigm.
Citation: Wittevrongel B, Van Hulle MM (2016) Frequency- and Phase Encoded SSVEP Using Spatiotemporal Beamforming. PLoS ONE 11(8): e0159988. https://doi.org/10.1371/journal.pone.0159988
Editor: Mikhail A. Lebedev, Duke University, UNITED STATES
Received: December 17, 2015; Accepted: June 3, 2016; Published: August 3, 2016
Copyright: © 2016 Wittevrongel, Van Hulle. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The (anonymized) recordings and analysis code are available from Open Science Framework with following url: osf.io/35r92.
Funding: BW is supported by the Agency for Innovation by Science and Technology in Flanders (IWT). MMVH is supported by research grants received from the program Financing program (PFV/10/008) of the KU Leuven, the Belgian Fund for Scientific Research – Flanders (G088314N, G0A0914N), the Inter-university Attraction Poles Programme -- Belgian Science Policy (IUAP P7/11), the Flemish Regional Ministry of Education (Belgium) (GOA 10/019), and the Hercules Foundation (AKUL 043).
Competing interests: The authors have declared that no competing interests exist.
Introduction
The steady-state visual evoked potential (SSVEP) is a neurophysiological response to a periodic visual stimulus commonly gauged with electroencephalography (EEG) over the occipital cortex when using stimuli flickering at a frequency above 6 Hz but in practice lower than about 30 Hz (due to the usual 60 Hz screen refresh rate and the EEG bandwidth). The recorded EEG signals not only resonate at that frequency but also contain a number of harmonics. When considering a display with several disjoint, spatially delimited stimuli, each one flickering at a different frequency, the stimulus the subject is currently gazing at can be inferred from a spectral analysis of the recorded EEG signals. This is also the principle behind the SSVEP-based brain-computer interface (BCI) where those stimuli become selectable targets in a subject interaction paradigm. However, a simple frequency analysis technique based on the (fast) Fourier transform [1, 2] typically requires long (i.e., 3 seconds or more [3, 4]) signals to accurately discriminate targets flickering at nearby frequencies. Further studies have led to several SSVEP detection techniques that are able to work with shorter signals such as Similarity of Background (SOB) [5], Minimum Energy Combination (MEC) [6], time-domain analysis [7, 8], and the widely adopted Canonical Correlation Analysis (CCA) [9–12] and its variants [13, 14].
The number of frequency-coded targets is not only limited by the harmonics of the stimulus frequency, but also by the screen refresh rate: when using a 50% duty-cycle stimulation (i.e., ‘on-off’ stimulation) the usable frequencies are restricted to integer dividers of the screen refresh rate. This restriction was demoted by replacing on-off stimulation with screen luminosity modulation [15].
As an alternative to frequency coding, also phase coding of the targets has been suggested [15–21]: (a subset of) targets flicker at the same frequency but with different phase ‘lags’. However, discriminating phases is more challenging and, typically, the number of useable phases is also quite limited, especially when based on short signals [20].
Another attempt to increase the number of selectable targets is to consider for each target a unique combination of frequency and phase [22–24]. For the joint detection of frequency and phase, the CCA method was extended and shown to be useful for a high-speed BCI application [24]. We will also consider combined frequency/phase coded targets but propose a new decoding approach: spatiotemporal beamforming in combination with time-domain analysis of EEG signals. The beamformer was originally formulated as a spatial filter for radar, sonar and seismic data analysis [25]. It was also employed in EEG analysis to isolate the signal originating from a predefined brain location [26], to estimate the amplitude of an ERP component [27, 28], and to build a BCI application based on imagined movement detection [29]. Here we extend the beamformer to a spatiotemporal filter for combined frequency-phase SSVEP BCI.
Methods
Subjects
We recruited 21 subjects, for our experiments, 14 female and 7 male (average age 22.7 years, between 19 to 26 years). Prior to the experiment, and after being informed of its purpose and design, our subjects read and, when they agreed, signed an informed consent form previously approved by the ethical committee of our university hospital UZLeuven. All subjects had normal or corrected to normal vision and were paid for their participation. All experiments were done in a sound-attenuated, air-conditioned room.
Interface
The targets consisted of four identical squares sized 9.5 × 9.5 cm, horizontally and vertically separated by 5.4 cm gap (∼4.4°) and diagonally by a 7.6 cm gap (∼6.2°), projected on an LCD computer display (sized 24.1 inch, resolution 1920×1200, 60 Hz refresh rate). Each square was assigned a different combination of frequency (12 or 15 Hz) and phase (0 or π radians) (Fig 1). Subjects were seated in a comfortable arm chair at a distance of approximately 70 cm from the display.
Each of the four squares has a unique combination of frequency and phase.
The recording session consisted of 60 trials. At the beginning of each trial, one of the squares was cued (target) with its corresponding color while the other squares were shown in gray. The subject was asked to direct his/her gaze at the cued square during the entire five-second stimulation. The trial was initiated when the subject pressed a key on the keyboard after which the all the squares regained their color. During a trial, all squares were flickering, in accordance with their frequency/phase combination, by sinusoidally modulating their luminosities using the method described in [15]. Between trials, the subject was allowed to take a break.
Recording
EEG was recorded continuously (sampling speed 2048 Hz, common mode sense (CMS) referencing) using 32 active Ag/AgCl electrodes (BioSemi Active Two) placed according to the extended international 10-20 system. Additionally, six external electrodes were placed: two on the left and right mastoids, for further off-line re-referencing, and four around the eyes, one on the upper and lower side of the left eye (vertical), and one near the external canthus of each eye (horizontal), for electro-oculogram recording (EOG, bi-polar recording). Except for these external ones, all electrodes were mounted in the electrode cap that was placed on the subject’s head. Conductive gel was applied in each of the electrode holes, as well as on the surface of the external electrodes to reduce electrode impedance. Prior to the main experiment, an EOG calibration session was performed to offline remove eye movements and blinks, using the method described in Croft & Barry [30].
Processing
The recordings were re-referenced offline from a CMS reference to the average of both mastoid signals and the EOG signal used to remove eye artifacts following the method of Croft and co-workers [30]. The corrected signals were bandpass filtered between 5 and 20 Hz using a 4th order Butterworth filter and then cut into five-second epochs, time-locked to the stimulation onset. Finally, the epochs were downsampled to 512 Hz, and labeled according to the color cued square. For each subject, 60 five-second epochs were extracted and stored for further analysis.
Analysis
Time-domain approach.
In order to discriminate between the frequency-and-phase encoded targets, we adopt a time-domain approach [7, 8]: each epoch is cut into consecutive, non-overlapping and identical segments with lengths equal to a single period of stimulus frequency f (blue traces in Fig 2). As we are using frequencies of 12 and 15 Hz, we have two such lengths, 83 and 67 ms, respectively. Then, we compute averages (red traces in Fig 2) When the stimulus frequency of the target that is gazed at matches one of the assumed stimulus periods, the corresponding average trace will be periodic, else it will be (almost) flat.
Note that the actual analysis is done on a spatiotemporal epoch, involving multiple electrodes.
Beamforming.
The beamformer starts from an activation pattern (i.e., a template of the signal-of-interest) that represents the EEG response to a particular frequency-phase combination (see further on how activation patterns are constructed), and transforms it into a weight vector (i.e., a multivariate filter), so as to optimally isolate the the targeted response from noise and possible non-target related activity, by taking into account the information contained in the covariance matrix estimated from the entire data. As the signals-of-interest differ among targets, each target has its own template with length equal to a single period of the target’s stimulus frequency and thus also its own beamformer. The latter is applied then to an averaged segment of the same length. As the templates differ among targets, each target has its own template and thus also beamformer that is applied to an averaged segment with length equal to a single period of the target’s stimulus frequency. The following explanation therefore applies to each target individually. Let matrix be the spatiotemporal activation pattern of a given frequency-phase combination, where m represents the number of electrodes and n the number of samples. Note that n depends on the segment length (given the sampling rate which is fixed), and thus on the target frequency (but not its phase).
While the beamformer was originally formulated as a spatial filter, we extend it to a spatiotemporal filter. We adopt the formulation of van Vliet and coworkers [27]. A Linearly Constrained Minimum Variance (LCMV) spatial beamformer minimizes the variance of the beamformer output
:
(1)
where
is the spatial covariance matrix of the EEG segment
. By adding the linear constraint:
(2)
where
is the spatial activation pattern, we avoid trivial solutions of Eq (1), and signals that are similar to asp will be mapped to a value close to 1, allowing for an easy measure of similarity. The solution of Eq (1) under constraint Eq (2) can be found using the method of Lagrange multipliers [26]:
The spatial beamformer can be expanded to a spatiotemporal variant as follows. Let r be the number of segments, a matrix where each row is obtained by concatenating the rows of a corresponding segment Si (i = 1..r),
the covariance matrix of
, and
a vector containing the concatenated rows of A. The spatiotemporal LCMV beamformer
with the linear constraint
can now be calculated as:
(4)
and applied to the data as a simple weighted sum:
(5)
where
indicates the concatenated rows of a segment S.
Another way to obtain a spatiotemporal filter is to apply a spatial beamformer followed by a temporal beamformer
. Let
be the matrix containing the results of the spatial filter wsp applied to the EEG segments Si:
(6)
, the covariance matrix of
, together with a temporal activation pattern
, allows for a temporal beamformer to be obtained as follows:
(7)
By consecutively applying the beamformers constructed in Eqs (3) and (7) to the segment under consideration, a spatiotemporal beamformer is obtained:
(8)
We further refer to Eq (5) as the spatiotemporal beamformer (stBF) and to Eq (8) as the chained beamformer (chBF).
Activation Patterns.
Since each frequency-phase combination is expected to evoke a different EEG response, activation patterns need to be calculated for each target i (∈[1..4]). Let be the target epochs (i.e., the epochs during which the subject was focussing on target i). Each epoch in
is then cut into segments, using the time-domain approach described above, where frequency f is the stimulation frequency of target i. The spatiotemporal activation pattern Ast[i] is then calculated by averaging the obtained segments.
As spatial activation pattern for target i, the column of Ast[i] for which channel Oz reaches its maximal value is selected:
(9)
where
(10)
Channel Oz was chosen since it can be expected that the SSVEP response is strongest at this channel.
To calculate the temporal activation pattern at[i] for square i, all epochs in are spatially filtered using a beamformer that was constructed with asp[i] as activation pattern. The covariance matrix used for constructing this beamformer is also estimated on
. The spatially filtered target epochs are then cut into segments using the time-domain approach where the frequency f is set to the stimulation frequency of target i. The temporal activation pattern is given by the average of these segments. The pseudo-code of this algorithm is shown in Algorithm 1.
Algorithm 1: Calculate Activation Patterns
1: for i ≔ 1 .. 4 do
2: fi← stimulation frequency of target i
3: target epochs for target i
4: Cut epochs in into segments, using fi
5: Ast[i]← average segments
6: asp[i]← select column of Ast[i], using Eqs (9) and (10)
7: Build beamformer with asp[i] and
8: Filter all epochs in with beamformer
9: Cut filtered epochs into segments, using fi
10: at[i]← average segments
11: end for
Classifiers.
Training the beamformer-based classifier involves the construction of four beamformers and ‘training’ four binary (one-vs-all) sub-classifiers, one for each target. The activation patterns and the beamformers were constructed as described earlier in this section. The feature space of the sub-classifiers is given by the output of the beamformers and is therefore one-dimensional. A threshold is applied to this output in order to discriminate between the target-(positive class) and the non-target (negative class) epochs (binary classification). Using a stratified four-fold cross-validation (i.e., in each fold, all labels occur equally frequent) for the training data, we determined for each sub-classifier the optimal threshold by using a Receiver Operating Characteristic (ROC) analysis. (An ROC curve shows the binary classification performance when varying the threshold.) Since the maximum classification performance could be reached for multiple thresholds (equal ROC points or points on the maximal iso-performance line), we took the median of those thresholds as final threshold.
Classification of an epoch is performed by independently applying each sub-classifier (and its necessary preprocessing, i.e., cutting into segments) to obtain four binary predictions and scores. (Fig 2) The binary prediction is based on the threshold of the respective sub-classifier and the score is equal to the y-value after applying the respective beamformer. From the sub-classifiers that return a positive prediction, the one having the highest score is taken as winner. In case of no positive predictions, the winner is determined by the sub-classifier returning the highest score.
To estimate the performance of the beamformer-based classifier, we apply a stratified five-fold cross-validation scheme. Because of the relatively high inter-subject variability of EEG responses to SSVEP stimulation [31], the analysis is run for each subject separately.
We compare our beamformer to an extension of the popular Canonical Correlation Analysis (CCA) method [24], with the number of harmonics set to three. To the best of our knowledge, when using short time windows in a frequency-phase SSVEP setting, the most reliable classification performance has been reported for the extended CCA method.
Parameter sensitivity.
We assess the performance of the classifiers with different epoch lengths, from 250 ms to 3 s in steps of 250 ms. For each epoch length l, we select a window from 0.120 to 0.120+l seconds from the original five-second epochs, and run the analysis with these reduced epochs. With longer epoch lengths, more segments can be extracted, leading to more reliable and easier to discriminate averages (see also the discussion section). The first 100-150 ms of each trial corresponds to the latency of the brain to SSVEP stimulation. [32, 33] Including this in our analysis might lower performance, hence, we chose to start from 120 ms post stimulus-onset. Note that this was also used in the Nakanishi study we consider in our comparison [24]
Since SSVEP is a visual stimulation paradigm, it can be expected that primarily electrodes over the occipital pole are relevant. However, including channels surrounding the occipital cortex might provide information about noise sources and thus lead to a more accurate filter [34]. On the other hand, a smaller number of channels reduces the dimensionality of the spatiotemporal filter, which is expected to increase performance (less parameters to be estimated, cf., curse of dimensionality [35, 36]). To test the influence of the number of channels, we additionally run the analysis using a reduced electrode setup, one with 10 channels (Chenv = {Fz, Cz, Pz, P3, P4, PO3, PO4, O1, Oz and O2}) and one with only three occipital channels (Chocc = {O1, Oz and O2}). The complete 32-channel setup will be referred to as Chfull.
In addition to the sampling rate of 512 Hz, we also ran the analysis using further downsampled epochs, and assessed their influence on the prediction performance. Lower sampling rates lead to less samples per segment and, thus, also to a lower dimensionality of the covariance matrices used when constructing the beamformers. The downsampling rates tested are 256 and 128 Hz, and the number of samples per segment are summarized in Table 1.
T corresponds to one period of a given frequency. Segments are input to the beamformers.
Results
Activation Patterns
The activation patterns for one subject are shown in Fig 3. The spatial activation patterns confirm that occipital electrodes are most informative, and that parietal and parietal-occipital electrodes also can contribute to the SSVEP detection.
Each row corresponds to a different target. The left column shows the spatiotemporal activation patterns Ast[i] that are used for building spatiotemporal beamformers. The middle and right columns show the spatial (asp[i]) and temporal activation patterns (at[i]) that are employed to build the chained beamformer.
When comparing the temporal patterns of all subjects for a given target (Fig 4), it is clear that they are very subject dependent. Hence, activation patterns should be determined for each subject individually rather than for a population if one is not willing to sacrifice classification performance. Therefore, we will build subject-specific beamformers and also assess their classification performance in this way.
Each trace corresponds to the temporal activation pattern of one subject.
Classification
Fig 5 shows the performance of the beamformers for the three channel sets, and S1 and S2 Tables summarize the statistical significance. The chained beamformer performs best using Chenv on short epoch lengths (max. 1 s). Even though the differences are not statistically significant, Chocc has a lower median performance for almost all trial lengths. With stBF, both reduced channel sets significantly outperform the full channel set for epochs shorter than 2.25 seconds. Chenv reaches a median accuracy of 100% for 1.25 s epoch lengths, while Chocc needs an additional 0.5 seconds to achieve similar (median) performance. Only with a very small epoch length (250 ms), Chenv performs significantly worse than Chocc.
(A) chained beamformer, (B) spatiotemporal beamformer. Stars indicate significance levels based on Wilcoxon Rank Sum Test: *(p < 0.05), **(p < 0.01), ***(p < 0.001).
For Chenv, the effect of downsampling is shown in Fig 6 and S3 and S4 Tables. For both beamformers, there are only minor effects. None of the differences with the chBF are significant. The only significant differences with stBF are for an epoch length of 250 ms. However, in that case the median performance does not reach 70%, which is often taken as the minimum for reliable communication [37–40].
(A) chained beamformer, (B) spatiotemporal beamformer, both for channel set Chenv.
To compare the beamformers with the extended CCA method, a sampling rate of 512Hz and Chenv was used. Fig 7 shows the prediction performances, and the statistical significance is given in Classification. Except for the shortest epoch length, our beamformer-based classifiers outperform the CCA alternative. The stBF-variant is significantly better than the CCA-alternative for short epoch lengths (≤1.25 s). It is also clear that the variance of both beamformers is smaller than that of the CCA alternative. Note that the performance of the extended CCA method on our data is comparable to the original study.
Channel set Chenv and no downsampling was used. Stars indicate significance levels, based on a Wilcoxon Rank Sum Test: *(p < 0.05), **(p < 0.01), ***(p < 0.001).
The accuracies achieved by the four sub-classifiers during the analysis of the performance reported in Fig 7 are shown in S1 and S2 Figs, for the stBF- and chBF-based classifier respectively. For all targets, the sub-classifier accuracies are approximately equal. However, target 4 (i.e., 15Hz and π rad) has consistently the lowest detection accuracy for both beamformers.
Finally, Fig 8 shows two confusion matrices of the stBF classifier of Fig 7 for epoch lengths of 0.5 and 1 second. We observe that for target 4 there are more misclassifications (last column in each matrix) than for the other targets and that there are less misclassifications between targets of opposite phase than between targets of different frequencies (e.g., in Fig 8(A), compare predicted target 1 vs. actual target 4 (1.67%) and vs. actual target 2 (2.33%)).
The numbers on the diagonal indicate the predictive accuracy (in %) of the stBF-classifier for each target. The values displayed are the average of the confusion matrices of all subjects. (A) For an epoch length of 0.5 seconds. (B) For an epoch length of 1 second. Targets 1 and 4 flicker at 12 Hz but with opposite phase, targets 2 and 3 flicker at 15 Hz but with opposite phase.
Discussion
Effect of combining frequency and phase encoding on SSVEP
Generally, the SSVEP signal has an oscillatory waveform with the same temporal frequency and phase as the flickering stimulus and its harmonics [41] and is considered to be evoked by the electrophysiological activity of large synchronized neuronal populations in early visual cortex [42]. The feasibility of identifying targets from spatially disparate (in visual angle or screen coordinates) flickering stimuli is supported by several lines of evidence. Firstly, as the gaze is directed, the attended stimulus is present in the fovea whereas the unattended targets are in the periphery of the subject’s visual field. As early visual areas are retinotopically organized, disparate visual stimuli activate disparate anatomical locations and vice-versa, as clearly evidenced by several neuroimaging studies [43–45]. When using neuroimaging techniques or EEG in combination with inverse mapping techniques, it was shown that flickering stimuli at different locations in the visual field activate different SSVEP generators in the early visual system [46].
A second aspect is the cortical magnification effect: a disproportionately large amount of visual cortex is allotted to processing foveal input compared to more peripheral input [47, 48]. As a result, the electrophysiological response evoked by a stimulus increases when becoming closer to the center of the visual field [48]. This implies that the amplitude of a foveated flickering stimulus is not going to be influenced by an another flickering stimulus when the visual angle that separates them is large enough [49]. When flickering targets are in close proximity, modulatory effects in the recorded EEG signal can be expected, in addition to an effect of attention. This has been studied by Fuchs and co-workers [50]. When the subject is focusing and attending one flickering target, and ignoring the other one, and when the visual angle that separates these targets is smaller than 4°, then the SSVEP amplitude of the attended target is unaltered with a small SSVEP amplitude of the unattended target added to it (co-amplification). Hence, when the flickering frequencies of those two targets are identical, but their phases in opposition (180° phase difference), and their separating angles much smaller than 4°, then the SSVEP amplitude of the attended target will be demoted or even largely cancelled. In our case, te SSVEP responses to attended targets are not canceled by targets with identical frequencies but opposite phases (target 1 vs. 4 and target 2 vs. 3) because their separating angles are much larger (∼6.2°). On the other hand, as the separating angles between targets flickering with different frequencies is much smaller (∼4.4°), we observe a co-amplification effect in the EEG responses to the attended targets (Fig 9). This is also reflected in the confusion matrices (Fig 8) by a much larger degree of misclassification between targets with different flickering frequencies than with opposite phases.
EEG frequency spectrum, averaged over all subjects, revealing co-amplification effect: the peripheral target flickering at a different frequency (∼4.4° separating angle) adds a small amplitude to the SSVEP signal of the attended target. (A) For targets 1 and 2, flickering at 12 and 15 Hz, respectively. (B) For targets 3 and 4, flickering at 15 and 12 Hz, respectively.
Activation Patterns
The activation patterns confirm that the SSVEP response is mostly present in the occipital regions, however using only those channels does not yield maximum performance. While surrounding channels have a smaller SSVEP response, they do contain information about noise sources that might contaminate the occipital signals [34]. By taking these channels into account, performance improves. However, increasing the number of channels also increases the dimensionality of the beamformer and the size of the (training) data set needed to accurately estimate the covariance matrix. [35, 36]
The beamformer approach developed in this study constructs an activation pattern for each individual subject, which calls for a training session for each subject. Indeed, the temporal responses to phase-based SSVEP stimulation were observed to vary considerably among subjects (Fig 4), hindering the construction of one activation pattern for the entire population. On the other hand, when individuating the beamformer, it can take into account subject- and session-dependent noise sources, which has a positive effect on performance. In addition, when comparing our beamformers with extended CCA, we observed for the beamformer-based classifiers not only a better performance but also a smaller interquartile range (IQR) over the subject population. This indicates that, although they are developed per subject, the beamformer-based classifiers yield a more stable performance across subjects. While both the beamformer and the extended CCA approach use prior information (targets, subjects), the beamformer takes into account the possible presence of non-target activity and structured noise embedded in the trials by using the covariance matrix, whereas the CCA method only extracts target trials to estimate canonical correlations. We therefore advocate that the beamformer is more efficient in extracting relevant information, leading to a better performance.
Frequency resolution
Our time-domain analysis of SSVEP signals is influenced by the frequencies used for stimulating the targets. If the difference in segment lengths between targets increases then less segments will be needed to accurately discriminate between a waveform and a ‘flat’ response. Fig 10 shows the effect of epoch lengths for a 15 Hz SSVEP response at channel Oz when assuming targets with frequencies (12, 13, 14, and 15 Hz). It is clear that longer epoch lengths are needed when discriminating targets with similar stimulation frequencies. Indeed, for 250 ms, one can easily exclude the 12 Hz response from the list of possible target responses, but longer segments are needed to discriminate between the 14 and 15 Hz responses.
A signal from channel Oz in response to a 15 Hz stimulus, and its segments with lengths corresponding to one period (T) of 12, 13, 14 and 15 Hz. Signal length are (A) 250 ms, (B) 500 ms, (C) 750 ms and (D) 1000 ms.
Conclusion
In this study, we have introduced beamforming as a new method for simultaneous frequency and phase detection in SSVEP-based BCI application. We have shown that our method outperforms an extension of CCA which has been touted as the best method for targets encoded by individual frequency-phase combinations, even for 1.25s epoch lengths.
Supporting Information
S1 Table. P-values for the performance differences of the chBF-based classifier, using different channel sets.
Values are calculated using a two-sided Wilcoxon Rank-Sum Test. Significant values are indicates in bold.
https://doi.org/10.1371/journal.pone.0159988.s001
(PDF)
S2 Table. P-values for the performance differences of the stBF-based classifier, using different channel sets.
Values are calculated using a two-sided Wilcoxon Rank-Sum Test. Significant values are indicates in bold.
https://doi.org/10.1371/journal.pone.0159988.s002
(PDF)
S3 Table. P-values for the performance differences of the chBF-based classifier, using different downsampling rates.
Values are calculated using a two-sided Wilcoxon Rank-Sum Test. Significant values are indicates in bold.
https://doi.org/10.1371/journal.pone.0159988.s003
(PDF)
S4 Table. P-values for the performance differences of the stBF-based classifier, using different downsampling rates.
Values are calculated using a two-sided Wilcoxon Rank-Sum Test. Significant values are indicates in bold.
https://doi.org/10.1371/journal.pone.0159988.s004
(PDF)
S5 Table. P-values for the performance differences between classifiers.
A downsampling rate of 512 Hz and channel set Chenv was used. Values are calculated using a two-sided Wilcoxon Rank-Sum Test. Significant values are indicates in bold.
https://doi.org/10.1371/journal.pone.0159988.s005
(PDF)
S1 Fig. Accuracies of the four sub-classifiers based on stBF.
A downsampling rate of 512 Hz and channel set Chenv was used.
https://doi.org/10.1371/journal.pone.0159988.s006
(PDF)
S2 Fig. Accuracies of the four sub-classifiers based on chBF.
A downsampling rate of 512 Hz and channel set Chenv was used.
https://doi.org/10.1371/journal.pone.0159988.s007
(PDF)
Author Contributions
- Conceived and designed the experiments: BW MMVH.
- Performed the experiments: BW.
- Analyzed the data: BW.
- Contributed reagents/materials/analysis tools: BW.
- Wrote the paper: BW MMVH.
References
- 1. Cheng M, Gao X, Gao S, Xu D. Design and implementation of a brain-computer interface with high transfer rates. Biomedical Engineering, IEEE Transactions on. 2002;49(10):1181–1186.
- 2. Kelly SP, Lalor EC, Reilly RB, Foxe JJ. Visual spatial attention tracking using high-density SSVEP data for independent brain-computer communication. Neural Systems and Rehabilitation Engineering, IEEE Transactions on. 2005;13(2):172–178.
- 3. Wu CH, Chang HC, Lee PL, Li KS, Sie JJ, Sun CW, et al. Frequency recognition in an SSVEP-based brain computer interface using empirical mode decomposition and refined generalized zero-crossing. Journal of Neuroscience Methods. 2011;196(1):170–181. Available from: http://www.sciencedirect.com/science/article/pii/S0165027010007028. pmid:21194547
- 4. Liu Q, Chen K, Ai Q, Xie SQ. Review: recent development of signal processing algorithms for SSVEP-based brain computer interfaces. J Med Biol Eng. 2013;.
- 5. Wu Z. SSVEP Extraction Based on the Similarity of Background EEG. PloS one. 2014;9(4):e93884. pmid:24709951
- 6. Friman O, Volosyak I, Gräser A. Multiple channel detection of steady-state visual evoked potentials for brain-computer interfaces. Biomedical Engineering, IEEE Transactions on. 2007;54(4):742–750.
- 7. Luo A, Sullivan TJ. A user-friendly SSVEP-based brain–computer interface using a time-domain classifier. Journal of neural engineering. 2010;7(2):026010.
- 8.
Manyakov NV, Chumerin N, Combaz A, Robben A, Van Hulle MM. Decoding SSVEP Responses using Time Domain Classification. In: IJCCI (ICFC-ICNC); 2010. p. 376–380.
- 9. Lin Z, Zhang C, Wu W, Gao X. Frequency recognition based on canonical correlation analysis for SSVEP-based BCIs. Biomedical Engineering, IEEE Transactions on. 2006;53(12):2610–2614.
- 10. Bin G, Gao X, Yan Z, Hong B, Gao S. An online multi-channel SSVEP-based brain–computer interface using a canonical correlation analysis method. Journal of neural engineering. 2009;6(4):046002. pmid:19494422
- 11. Xie J, Xu G, Wang J, Zhang F, Zhang Y. Steady-State Motion Visual Evoked Potentials Produced by Oscillating Newton’s Rings: Implications for Brain-Computer Interfaces. PLoS ONE. 2012 06;7(6):e39707. Available from: http://dx.doi.org/10.1371%2Fjournal.pone.0039707. pmid:22724028
- 12. Yin E, Zhou Z, Jiang J, Yu Y, Hu D. A Dynamically Optimized SSVEP Brain-Computer Interface (BCI) Speller. Biomedical Engineering, IEEE Transactions on. 2015 June;62(6):1447–1456.
- 13. Zhang Y, Zhou G, Jin J, Wang X, Cichocki A. Frequency recognition in SSVEP-based BCI using multiset canonical correlation analysis. International journal of neural systems. 2014;24(04):1450013. pmid:24694168
- 14. Chen X, Wang Y, Gao S, Jung TP, Gao X. Filter bank canonical correlation analysis for implementing a high-speed SSVEP-based brain–computer interface. Journal of neural engineering. 2015;12(4):046008. pmid:26035476
- 15. Manyakov NV, Chumerin N, Robben A, Combaz A, van Vliet M, Van Hulle MM. Sampled sinusoidal stimulation profile and multichannel fuzzy logic classification for monitor-based phase-coded SSVEP brain–computer interfacing. Journal of neural engineering. 2013;10(3):036011. pmid:23594762
- 16. Lee PL, Sie JJ, Liu YJ, Wu CH, Lee MH, Shu CH, et al. An SSVEP-actuated brain computer interface using phase-tagged flickering sequences: a cursor system. Annals of Biomedical Engineering. 2010;38(7):2383–2397. pmid:20177780
- 17. Lopez-Gordo M, Prieto A, Pelayo F, Morillas C. Use of phase in brain–computer interfaces based on steady-state visual evoked potentials. Neural processing letters. 2010;32(1):1–9.
- 18.
Zhu D, Garcia-Molina G, Mihajlović V, Aarts RM. Online BCI implementation of high-frequency phase modulated visual stimuli. In: Universal Access in Human-Computer Interaction. Users Diversity. Springer; 2011. p. 645–654.
- 19. Falzon O, Camilleri K, Muscat J. Complex-valued spatial filters for SSVEP-based BCIs with phase coding. Biomedical Engineering, IEEE Transactions on. 2012;59(9):2486–2495.
- 20.
Manyakov N, Chumerin N, Combaz A, Robben A, van Vliet M, Van Hulle M. Decoding Phase-Based Information from Steady-State Visual Evoked Potentials with Use of Complex-Valued Neural Network. In: Yin H, Wang W, Rayward-Smith V, editors. Intelligent Data Engineering and Automated Learning—IDEAL 2011. vol. 6936 of Lecture Notes in Computer Science. Springer Berlin Heidelberg; 2011. p. 135–143. Available from: http://dx.doi.org/10.1007/978-3-642-23878-9_17.
- 21. Manyakov NV, Chumerin N, Van Hulle MM. Multichannel decoding for phase-coded SSVEP brain–computer interface. International journal of neural systems. 2012;22(05):1250022. pmid:22963395
- 22. Jia C, Gao X, Hong B, Gao S. Frequency and phase mixed coding in SSVEP-based brain–computer interface. Biomedical Engineering, IEEE Transactions on. 2011;58(1):200–206.
- 23.
Chen X, Wang Y, Nakanishi M, Jung TP, Gao X. Hybrid frequency and phase coding for a high-speed SSVEP-based BCI speller. In: Engineering in Medicine and Biology Society (EMBC), 2014 36th Annual International Conference of the IEEE. IEEE; 2014. p. 3993–3996.
- 24. Nakanishi M, Wang Y, Wang YT, Mitsukura Y, Jung TP. A high-speed brain speller using steady-state visual evoked potentials. International journal of neural systems. 2014;24(06):1450019. pmid:25081427
- 25. Van Veen BD, Buckley KM. Beamforming: A versatile approach to spatial filtering. IEEE assp magazine. 1988;5(2):4–24.
- 26. Van Veen BD, Van Drongelen W, Yuchtman M, Suzuki A. Localization of brain electrical activity via linearly constrained minimum variance spatial filtering. Biomedical Engineering, IEEE Transactions on. 1997;44(9):867–880.
- 27. van Vliet M, Chumerin N, De Deyne S, Wiersema JR, Fias W, Storms G, et al. Single-trial ERP component analysis using a spatio-temporal LCMV beamformer. Biomedical Engineering, IEEE Transactions on. 2015;PP(99):1–1.
- 28. Wittevrongel B, Van Hulle MM. Faster P300 Classifier Training Using Spatiotemporal Beamforming. International Journal of Neural Systems. 2016;26(3):1650014. pmid:26971787
- 29. Grosse-Wentrup M, Liefhold C, Gramann K, Buss M. Beamforming in noninvasive brain–computer interfaces. Biomedical Engineering, IEEE Transactions on. 2009;56(4):1209–1219.
- 30. Croft R, Barry R. Removal of ocular artifact from the EEG: a review. Neurophysiologie Clinique/Clinical Neurophysiology. 2000;30(1):5–19. pmid:10740792
- 31. Tobimatsu S, Tomoda H, Kato M. Normal variability of the amplitude and phase of steady-state VEPs. Electroencephalography and Clinical Neurophysiology/Evoked Potentials Section. 1996;100(3):171–176.
- 32. Di Russo F, Spinelli D. Electrophysiological evidence for an early attentional mechanism in visual processing in humans. Vision research. 1999;39(18):2975–2985. pmid:10664797
- 33.
Di Russo F, Teder-Sälejärvi WA, Hillyard SA. Steady-state VEP and attentional visual processing. The cognitive electrophysiology of mind and brain (Zani A, Proverbio AM, eds). 2002;p. 259–274.
- 34. Haufe S, Meinecke F, Görgen K, Dähne S, Haynes JD, Blankertz B, et al. On the interpretation of weight vectors of linear models in multivariate neuroimaging. Neuroimage. 2014;87:96–110. pmid:24239590
- 35.
Pruzek RM. High Dimensional Covariance Estimation: Avoiding the ‘Curse of Dimensionality’. In: Proceedings of the First US/Japan Conference on the Frontiers of Statistical Modeling: An Informational Approach. Springer; 1994. p. 233–253.
- 36.
Schoukens J, Pintelon R. Identification of linear systems: a practical guideline to accurate modeling. Elsevier; 2014.
- 37. Combaz A, Chatelle C, Robben A, Vanhoof G, Goeleven A, Thijs V, et al. A comparison of two spelling brain-computer interfaces based on visual P3 and SSVEP in Locked-In Syndrome. PloS one. 2013;8(9):e73691. pmid:24086289
- 38. Brunner C, Allison B, Altstätter C, Neuper C. A comparison of three brain-computer interfaces based on event-related desynchronization, steady state visual evoked potentials, or a hybrid approach using both signals. Journal of neural engineering. 2011;8(2):025010. pmid:21436538
- 39. Kübler A, Neumann N, Wilhelm B, Hinterberger T, Birbaumer N. Predictability of brain-computer communication. Journal of Psychophysiology. 2004;18(2/3):121–129.
- 40. Kübler A, Birbaumer N. Brain-computer interfaces and communication in paralysis: Extinction of goal directed thinking in completely paralysed patients? Clinical neurophysiology. 2008;119(11):2658–2666. pmid:18824406
- 41.
Regan D. Human brain electrophysiology: evoked potentials and evoked magnetic fields in science and medicine. Elsevier; 1989.
- 42. Rager G, Singer W. The response of cat visual cortex to flicker stimuli of variable frequency. European Journal of Neuroscience. 1998;10(5):1856–1877. pmid:9751156
- 43. Bridge H. Mapping the visual brain: how and why. Eye. 2011;25(3):291–296. pmid:21102491
- 44. Vanegas MI, Blangero A, Kelly SP. Exploiting individual primary visual cortex geometry to boost steady state visual evoked potentials. Journal of neural engineering. 2013;10(3):036003. pmid:23548662
- 45. Raz N, Levin N. Cortical and white matter mapping in the visual system-more than meets the eye: on the importance of functional imaging to understand visual system pathologies. Neurovision: Neural bases of binocular vision and coordination and their implications in visual training programs. 2015;p. 90.
- 46. Di Russo F, Pitzalis S, Aprile T, Spitoni G, Patria F, Stella A, et al. Spatiotemporal analysis of the cortical sources of the steady-state visual evoked potential. Human brain mapping. 2007;28(4):323–334. pmid:16779799
- 47. Cowey A, Rolls E. Human cortical magnification factor and its relation to visual acuity. Experimental Brain Research. 1974;21(5):447–454. pmid:4442497
- 48. Sutter EE. The brain response interface: communication through visually-induced electrical brain responses. Journal of Microcomputer Applications. 1992;15(1):31–45.
- 49. Zhao J, Zhang Z, Zhang C, Tang Y, Liu Z. Neural suppression of distractors surrounding the spotlight: Evidence from steady-state visual evoked potentials. Chinese Science Bulletin. 2012;57(14):1680–1684.
- 50. Fuchs S, Andersen SK, Gruber T, Müller MM. Attentional bias of competitive interactions in neuronal networks of early visual processing in the human brain. NeuroImage. 2008;41(3):1086–1101. pmid:18424083