Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Method for enhancing single-trial P300 detection by introducing the complexity degree of image information in rapid serial visual presentation tasks

  • Zhimin Lin,

    Roles Data curation, Validation, Writing – original draft

    Affiliation China National Digital Switching System Engineering and Technological Research Center, Zhengzhou, China

  • Ying Zeng,

    Roles Project administration, Validation, Writing – review & editing

    Affiliations China National Digital Switching System Engineering and Technological Research Center, Zhengzhou, China, Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China

  • Li Tong,

    Roles Conceptualization, Formal analysis, Project administration, Supervision

    Affiliation China National Digital Switching System Engineering and Technological Research Center, Zhengzhou, China

  • Hangming Zhang,

    Roles Supervision

    Affiliation China National Digital Switching System Engineering and Technological Research Center, Zhengzhou, China

  • Chi Zhang,

    Roles Conceptualization, Supervision

    Affiliation China National Digital Switching System Engineering and Technological Research Center, Zhengzhou, China

  • Bin Yan

    Roles Conceptualization, Funding acquisition, Project administration, Supervision

    ybspace@hotmail.com

    Affiliation China National Digital Switching System Engineering and Technological Research Center, Zhengzhou, China

Abstract

The application of electroencephalogram (EEG) generated by human viewing images is a new thrust in image retrieval technology. A P300 component in the EEG is induced when the subjects see their point of interest in a target image under the rapid serial visual presentation (RSVP) experimental paradigm. We detected the single-trial P300 component to determine whether a subject was interested in an image. In practice, the latency and amplitude of the P300 component may vary in relation to different experimental parameters, such as target probability and stimulus semantics. Thus, we proposed a novel method, Target Recognition using Image Complexity Priori (TRICP) algorithm, in which the image information is introduced in the calculation of the interest score in the RSVP paradigm. The method combines information from the image and EEG to enhance the accuracy of single-trial P300 detection on the basis of traditional single-trial P300 detection algorithm. We defined an image complexity parameter based on the features of the different layers of a convolution neural network (CNN). We used the TRICP algorithm to compute for the complexity of an image to quantify the effect of different complexity images on the P300 components and training specialty classifier according to the image complexity. We compared TRICP with the HDCA algorithm. Results show that TRICP is significantly higher than the HDCA algorithm (Wilcoxon Sign Rank Test, p<0.05). Thus, the proposed method can be used in other and visual task-related single-trial event-related potential detection.

Introduction

The increasing demand for computer images and storage has resulted in abundant image data. Computer vision (CV) plays a remarkable role in current image retrieval because of its increasing computer processing speed. Although CV has been successfully applied in image retrieval, these achievements are limited to special conditions. The effective presentation for the interested image is difficult in image retrieval. Human vision (HV) is superior to CV in terms of its robust and general purpose image recognition ability. HV can also easily recognize target images with large variations. Moreover, HV processing time on a recognition task can be as fast as a few milliseconds, because event-related potential (ERP) has rapid specific response after onset of stimulus [1]. A brain–computer interface (BCI) is a state-of-the-art human–machine interaction technology [2, 3]. This interface records signals of human brain activity (e.g., electroencephalogram) to analyze human intention and then sends the results to a computer. The P300 component in an EEG signal can be used as an indicator to categorize the interest of a user [4]. Moreover, the P300 is a common ERP component and shows a peak waveform when small probability events are observed after approximately 300–500 ms [5]. And many scholars are build a variety of BCI system using the P300 and other EEG components. Erwei Yin et al. combine the P300 component and steady-state visually evoked potential (SSVEP) to build a high-performance hybrid BCI speller system[68]. Dewen Hu et al. used the P300 component to construct an auditory / tactile visual saccade-independent P300 BCI system[9]. These P300 components can be exploited when building a target-image detector based on rapid serial visual presentation (RSVP) paradigm [10, 11].

The P300 component exhibits significant waveform characteristics in the time domain. Thus, P300 can be extracted through average multiple trials of EEG signals. In particular, some parameters of P300 component, such as latency and amplitude, are not fixed, but these parameters are important in evaluating P300. In [12], the latency and amplitude of the ERP may vary over time for a given task in relation to different experimental parameters, such as target probability and stimulus meaning. Parra et al. proposed the hierarchical discriminant component analysis (HDCA) algorithm [1316] to overcome the temporal variability of latency and amplitude. This group separated single-trial EEG signals into several time windows and calculated the spatial filter to maximize the separation between target and nontarget categories. Alpert et al. proposed the hierarchical discriminant principal component analysis (HDPCA) algorithm[17], which introduces the principal component analysis for dimensionality reduction. Marathe et al. developed the sliding HDCA (sHDCA) algorithm [18, 19]. These methods are often focused on the EEG aspects.

Several scholars have considered the combination of EEG and CV to enhance the recognition accuracy. Sajda et al. proposed a closing the loop in cortically-coupled CV (close-loop 3CVision) system to detect the category of subject’s interest image [13, 16]. In their system, the score of the subject interest in an image was estimated by the HDCA algorithm [14], and the combination of this score and CV infers the interest image category from a large database. Wang et al. proposed a similar closed-loop system for face retrieval by coupling EEG-based target image labeling and CV-based label propagation [20]. These techniques involve decision-level fusion that first calculates the EEG interest score, and then CV combines the interest score to guess the user’s target of interest. We believe that the CV can be further integrated in the calculation of the interest score to obtain better results.

We propose a novel method for target recognition in which we can acquire a priori estimate of the deformation of the P300 component of the target image by estimating image complexity (IC) and train the classifier to improve the overall performance. Early studies [5, 12, 1719, 21, 22] have shown that the specific content of an image will affect the P300 component amplitude and latency. In this study, we used the deep neural network (DNN) to extract image semantic and pixel information [23, 24] and quantify the IC. Moreover, we trained the classifiers separately according to the different complexity ranges. During testing, we used all classifiers and synthesized the results to arrive at a final score. We called this process as a priori image recognition algorithm of IC, that is, Target Recognition using Image Complexity Priori (TRICP) algorithm. We compared TRICP with HDCA algorithm under different classifier parameters, and the results show that TRICP is significantly higher than the HDCA algorithm (Wilcoxon Sign Rank Test, p<0.05).

Methods

Participants

A total of 19 subjects (16 males and 3 females, age range of 21 to 24, and right-handed) participated in the experiment. All subjects were students of Zhengzhou University and did not have any previous training in the task, and all participants were recruited in January 2016. The subjects exhibited normal or corrected-to-normal vision with no neurological problems and were financially compensated for their participation. This study was conducted after we obtained informed consent and Ethics Committee approval of China National Digital Switching System Engineering and Technological Research Center. All of the participants provided their written informed consent to participate in this study.

Visual stimuli and procedure

The participants were seated 75 cm in front of a monitor. Images were chosen from the Caltech-256 database [25] and presented to the subjects using the RSVP paradigm [10, 26]. The images were shown in blocks of 96 and flashed at 5 Hz (Fig 1). Each image was positioned at the center of the computer monitor. A fixation cross was flashed immediately prior to the presentation of each block to allow the users to focus their gaze on the images during the RSVP sequences. For these tasks, the RSVP sequence consisted of 25 blocks (a total of 2400 images, i.e., 300 target images from 25 categories and 2100 non-target images from 175 categories). Each block consisted of 12 target images from one category and 84 non-target images from seven categories (12 images in each category). The target categories for each block differed from one another and are shown in Table 1.

thumbnail
Fig 1. Rapid serial visual presentation (RSVP) paradigm.

The RSVP sequence consisted of 25 blocks (a total of 2400 images, i.e., 25 target categories with 300 images and 175 nontarget categories with 2100 images), and each block comprised one target category (12 images) and seven nontarget categories (84 images). The images are presented in 25 blocks, with a distinct target category in each block. The target categories of each block are shown in Table 1. Each image is presented for 200 ms (The image is similar but not identical to the original image, and is therefore for illustrative purposes only).

https://doi.org/10.1371/journal.pone.0184713.g001

System overview

In this paper, we propose a TRICP method for image retrieval (Fig 2). The algorithm includes three major components, namely, CV, EEG, and mix modules. First, we used the CV module to estimate IC. We sorted the IC score and divided the images into three categories, namely, high-, medium-, and low-complexity images. Then, we recoded all image EEG data and trained three corresponding EEG classifiers (high-, medium-, and low-complexity classifier), on the data sets. Finally, during testing, we presented a picture to the participants and recoded the EEG signal and estimated IC. We calculated the EEG scores using the three classifiers and combined the three scores and the IC for a final score using a set of weights. We determined the category according to the final score.

thumbnail
Fig 2. System overview.

First, the sample data are divided into equal three parts according to the image complexity (IC). The EEG data induced by high-, medium-, and low-complexity images. We trained the classifiers separately on the different data sets. During testing, we first determined the complexity and category (high-, medium-, or low-complexity image) of the test picture. Then, we calculated the interest scores of the EEG induced by a test image using the corresponding classifier. The result was combined with a certain weight to obtain the final decision score (The image is similar but not identical to the original image, and is therefore for illustrative purposes only).

https://doi.org/10.1371/journal.pone.0184713.g002

1) EEG module.

EEG data were acquired by a g.USBamp system (G.Tec company) using 16 electrodes distributed in accordance with the international 10–20 system. The EEG data were sampled at 2400 Hz using 200 Hz low-pass and 50 Hz notch filters. Prior to scoring the images, we pre-processed the EEG data through the following steps: downsampling to 600 Hz, band-pass filtering (0.1–60 Hz) with a 10th order butterworth filter, baseline correction, and ocular artifact reduction. Here, zero- delay filtering was implemented using the filtfilt() function in MATLAB. Afterward, the EEG data were divided into epochs. Each epoch consisted of 1000 ms of EEG data after the stimulus onset.

Analysis of the ERP using HDCA algorithm was performed as described by Parra et al [13, 14, 26]. The HDCA algorithm can be divided into two layers. First, the HDCA algorithm was employed to obtain the average data and divide the original EEG data by time window size. The weight of each channel was then calculated in each time window to maximize the differences between the target and nontarget classes. In our study, the time window size cannot be determined in advance. Thus, we chose 25 ms as the time window size after numerous experimental repetitions. The weight of each channel in each time window was calculated by Fisher linear discriminant (FLD). In each time window, the EEG signal was reduced to one dimension, such as in Eq (1), as follows: (1) where xi[(k−1)N+n] represents the kth separate time-window value from the single-trial data. The variable corresponds to the EEG activity at the data sample point n measured by electrode i. w is a set of spatial weights. Weight vector wki is found for the kth window and i electrode following each image presentation (T is the temporal resolution of the time window, N is the sampling time point of the time window, FS is the sampling rate, K is the number of time window, and n = 1,2,⋯,N, N = T/FS, 0 ≤ kK).

(2)

The results for the separate time windows (yk) are then combined in a weighted yk average to provide a final interest score (yIS) for each image. FLD analysis was employed to calculate the spatial coefficient wki, and logistic regression was adopted to calculate for the temporal coefficient vk. We specified a threshold greater than the threshold value, that is, a target.

In this paper, the time windows of HDCA are adjustable parameters. In order to verify the effectiveness of this TRICP, we set the time windows to be 100ms, 50ms, 33ms and 25ms respectively, and we call it Classifier I, II, III, IV, respectively.

2) CV module.

The CV module ranks all images through the IC. We used IC to describe the brain processing efficiency of image information. In this paper, we assumed that the human brain processing complex image is higher than the simple image. Thus, the subjects’ EEG signals induced by complex and simple target images vary. The complex and simple images are nonobjective. We aimed to use the knowledge in the CV field to accurately quantify the IC. The convolution neural network (CNN) is the most effective image classifier, with its importance partly caused by its mechanism which mimics the human brain processing of an image. The CNN is a deep neural network. At the CNN bottom layer, the image features are represented by texture, edge, structure, and other characteristics. High-layer features of CNN are often a combination of underlying features, representing more abstract semantic features. We considered that if an image containing simple semantics in the high-level net mapping, the feature weight should be focused on individual features, and irrelevant feature weights are small. Additionally, a semantic complex image (containing more semantics) will have more feature weights that are larger in CNN high-layer mapping.

Thus, the complexity of the semantic level can be described through the high-layer feature weights of the CNN. These weights of a simple semantic image are more concentrated, whereas those of a complex semantic image are more dispersed. Similarly, the complexity of an image structure can be described through the underlying feature weight distribution of a CNN model.

Therefore, we extracted the feature vector of the image in a layer of the CNN and converted this vector into IC using the following formula: (3) where IC is the image complexity, and f is the feature weight vector of the image in some layer of CNN, fnum is the number of characteristic features, and k is a parameter used to distinguish differences between high and low ICs. It is worth noting that, when k is greater than 1, the image complexity ranking is the same. In this paper, k = 2. Early studies have shown that the P300 latency and amplitude caused by different semantics images will vary. Therefore, we believe that the P300 will vary because of the different ICs. Relative to the traditional machine vision for the definition of image complexity, the Eq 3 is special and it is a meaningful innovation. In this paper, we adopted the AlexNet network proposed in [27]. The AlexNet won achieved a winning top-5 test error rate of 15.3%, compared with 26.2% achieved by the second-best entry in the ILSVRC-2012 competition. We believe that AlexNet imitates the characteristics of the human visual system, and AlexNet can be used in our study reference to some extent. We used the model trained in the caffe framework [28]. The AlexNet network consists of eight layers, and we used the fifth layer feature to calculate IC in the following analysis. We believe that features of the middle layer is reasonable, it can better combine the semantic and structural information.

The Fig 3 shows the results of a group of images sorted by IC.

thumbnail
Fig 3. Three group images are sorted from high to low IC (The image is similar but not identical to the original image, and is therefore for illustrative purposes only).

https://doi.org/10.1371/journal.pone.0184713.g003

3) Mixed module.

An important innovation in this study is the introduction of image information (i.e., IC) into the calculation of the EEG final score. The IC value does not contain any subjective intent (interested or not interested) when the participants viewed the picture, therefore IC cannot be directly introduced into the interest score. We used a specific fusion method as follows: the training data set was divided into three parts according to the IC, and each part train a classifier alone. During testing, the parts of interest scores were fused according to the weight into the final interest score.

The training set was divided into three parts, namely, the EEG signals from high-, medium-, and low-complexity images. Then, the corresponding training classifiers were applied independently. During testing, we calculated three interest scores of the test image using the three classifiers corresponding to the IC. The final score was combined using the following rules: the classifier weight of the test image [high IC (HIC), medium IC (MIC), or low IC (LIC)] was given a weight of α, and other classifier weights were assigned to β. This process can be expressed by Eq (4). Different classifier weights can also be assigned through a more complex process according to the IC, such as using a linear classifier (SVM, Fisher, or logistic regression) on the validation set to compute for the different classifier weights for the final score. For convenience of expression, we used a simple method. (4) where T is an image, IC(T) is the IC of T computed by Eq (3). The IChigh_th and ICmid_th are the IC threshold values. High IC (HIC) is obtained when the IC(T) is greater than IChigh_th, whereas low IC (LIC) is characterized by IC(T) value less than ICmid_th. Values in between HIC and LIC define MIC. yIS_hclass(T), yIS_mclass(T), and yIS_lclass(T) are the interest scores computed using the high-, medium-, and low-complexity classifiers, respectively, according to the single-trial EEG induced by picture T. score(T) is the calculated final score combining the information from the EEG and the image. Here, α = 0.5, β = 0.25.

Evaluation of the algorithm performance

A five-fold cross validation was conducted to determine the accuracy of all classification algorithms applied to the EEG data. Data from each subject were divided into five equal-sized trial blocks. Classifiers were trained on four of the five blocks and then tested on the remaining block. This process was repeated five times, such that each of the five trial blocks was used once as an independent testing set. Each training block used to train a classifier was divided into two parts. Performance was evaluated based on the area under the receiver operating characteristic (ROC) curve (AUC) [29].

Results

A. Event-related responses (targets vs. nontargets)

We analyzed event-related responses to study the mean ERP, which was averaged over repeated trials under the same stimulus. Fig 4 depicts an ERP elicited by the target and nontarget ERPs at electrode Pz collapsed over blocks for a single sample subject. Fig 4 is consistent with literatures [19, 22]. Note that on an average, despite the rapid sequence of events and the overlapping responses, the main divergence between the target and nontarget ERPs occurs between 400–600 ms presentation. These results are consistent with the literature [5]. The same results can be observed with single-trial responses, as shown in Fig 4(A).

thumbnail
Fig 4. Target and non-target class waveforms.

(A) All single-trial event-related potentials (ERPs) to target images at electrode Pz. (B) All single-trial ERPs to nontarget images at electrode Pz. (C) Grand averages across all trials of the target EEG signals at electrode Pz. (D) Grand averages across all trials of the nontarget EEG signals at electrode Pz.

https://doi.org/10.1371/journal.pone.0184713.g004

In particular, each image differed in RSVP sequence and was presented for 200 ms. Hence, the participants focused at a 5 Hz stimulus source. The EEG signals generated a mixed 5 and 10 Hz harmonic (SSVEP).

B. Effect of IC on ERP

To study the effect of different ICs, we averaged all subject ERP waveforms of the same image (Fig 5). We sorted all images according to IC and defined the first one-third of images as HIC, while the middle one-third is defined as MIC, and the last one-third is defined as LIC. Fig 5(A) and 5(B) show the target and nontarget ERP data caused by HIC images from electrode Pz, whereas Fig 5(C) and 5(D) show those caused by MIC images. Additionally, Fig 5(E) and 5(F) show those data caused by LIC images. The red, blue, and green lines in Fig 5(G) are the averaged ERP waveforms of the trials shown in Fig 5(A), 5(C) and 5(E), respectively. Similarly, the red, blue, and green lines in Fig 5(H) are the averaged ERP waveforms of the trials shown in Fig 5(B), 5(D) and 5(F), respectively.

thumbnail
Fig 5. Effect of IC on ERP signal.

ERPs stimulated by the (A) high-complexity target image, (B) high-complexity nontarget image, (C) medium-complexity target image, (D) medium-complexity nontarget image, (E) low-complexity target image, and (F) low-complexity nontarget image. (G) Trial averaged target ERP wave forms calculated from (A), (C) and (E), where the red, blue, and green lines indicate the ERP components inspired by high-, medium- and low-complexity image. Similarly, images in (H) are indicative of nontarget image.

https://doi.org/10.1371/journal.pone.0184713.g005

Earlier studies have shown that information, such as the meaning of stimulus image, had an effect on P300 composition. Here, we illustrated the relationship through IC. Fig 5(G) shows that the amplitude of P300 excited by the HIC images was lower than that of the P300 excited by the MIC and LIC images, and the latency also varied. Fig 5(H) shows that the grand averaged ERPs of nontarget images did not significantly differ under the different IC conditions.

Table 2 shows the difference in the peak amplitudes and peak latencies of 19 participants under various IC conditions. The peak amplitude and latency were calculated using the maximum value of the averaged ERP of the different complexities. Table 2 shows that the amplitude and peak latency of HIC significantly differed from MIC and LIC (HIC, 4.76±1.09 μV; MIC, 5.44±0.9 μV; LIC, 5.49±1.22 μV; Wilcoxon Sign Rank Test, p<0.05). The amplitude of MIC did not significantly differ from that of LIC (p = 0.88). Table 2 shows that the peak latency of HIC significantly differed from those of MIC and LIC (HIC, 564.98±52.98 ms; MIC, 530.19±58.66 ms; LIC, 525.33±50.27 ms; Wilcoxon Sign Rank Test, p<0.05). The peak latency of MIC was not significantly differ from that of LIC (p = 0.52). On average, the P300 component induced by the HIC target image was 0.73 μV lower and the peak latency was delayed by 39.65 ms compared with the LIC image.

thumbnail
Table 2. The amplitude and latency changes at different complexities.

https://doi.org/10.1371/journal.pone.0184713.t002

Fali Li et al. research a relationships between the resting-state network and the P300, through a sample oddball cognitive task[30]. Fali Li et al. study indicated that P3 amplitude was significantly correlated with resting-state network topology, and no significant relationships were found for the corresponding P3 latency. However, the P300 component induced by the complex cognitive task is no more clearly conclusion. We calculated IC according to Eq (3) and the fifth layer features of the AlexNet network. We infer that the features of fifth layer showed better balance between the semantic and structural complexities. The results show that the P300 properties are different induced by different complexity images.

C. Single-trial detection

The AUC for each subject and the mean and SD for all subjects per algorithm group are shown in Table 3, Figs 6, 7, 8 and 9. In the experiment, each participant focused on different targets in diverse blocks (Table 1). The variation in specific meaning and complexity of the different target images led to changes in the latency and amplitude of the P300 component. The detection algorithms affected the precision of the single-trial P300, which was also demonstrated by Alpert et al.[17]. An interesting phenomenon is that TRICP achieves significantly better results than the HDCA algorithm in subjects with low AUC (e.g., Subjects 4, 11, 16). In subjects with higher AUC (e.g., Subjects 12, 18, 19), TRICP and HDCA algorithm results are similar. In subjects with moderate AUC, the TRICP better than the results of HDCA algorithm. This may indicate that, in less accurate subjects, the image is too complex may be an important reason.

thumbnail
Fig 6. Values of the areas under the receiver operating characteristic curve (AUC) of all subjects under the two algorithms, and the time windows is 100 ms.

https://doi.org/10.1371/journal.pone.0184713.g006

thumbnail
Fig 7. Values of the areas under the receiver operating characteristic curve (AUC) of all subjects under the two algorithms, and the time windows is 50 ms.

https://doi.org/10.1371/journal.pone.0184713.g007

thumbnail
Fig 8. Values of the areas under the receiver operating characteristic curve (AUC) of all subjects under the two algorithms, and the time windows is 33 ms.

https://doi.org/10.1371/journal.pone.0184713.g008

thumbnail
Fig 9. Values of the areas under the receiver operating characteristic curve (AUC) of all subjects under the two algorithms, and the time windows is 25 ms.

https://doi.org/10.1371/journal.pone.0184713.g009

To solve this problem, the thesis proposed the introduction of image information. The parameter which can predigest the deformation of ERP in accordance with the IC and then target training classifier was determined. Our proposed TRICP fusion method introduces the IC of an image on the basis of an existing algorithm and improves the accuracy of single-trial ERP detection.

Discussion

Current studies have shown that some deep neural networks process images are similar to the human brain. Agrawal et al. use the CNN based on the ImageNet image library to extract the features of the natural image, and use the middle layer as the image feature to train the fMRI visual coding model[31]. The results show that the visual coding model of CNN has achieved better prediction effect in the low-level visual area and the high-level visual area. Van Gerven et al. used a trained DNN to build a coding model to analyze the similarity between DNN and brain function brain area[32]. The experimental results show that the stimulus features exhibit hierarchical distribution on the deep neural network. Furtherly, Radoslaw et al. use a magnetoencephalography (MEG) and fMRI to observe the brain activity and compared with the DNN[33]. The results show that there is a corresponding relationship between the mapping of DNN low layer and high layer and the order of human brain vision signal processing, this together demonstrates the hierarchical structure similarity between DNN and human brain vision in spatial and temporal.

In this study, we extracted the features of an image through the different layers of the AlexNet network and converted these features to IC through Eq (3). The underlying features of the CNN network are more focused on the structural characteristics of the image. Thus, the construction can be considered as structural complexity. The features of high-level extraction of CNN network are more emphasized on the semantics of an image, such that the complexity can be regarded as semantic complexity. We believe that the characteristics of the middle layer is reasonable, it can better combine the semantic and structural information. Attention to complex images will require greater cognitive burden, and early studies have shown that the meaning of the stimulus image will affect the amplitude and peak latency of the P300 component. We carefully analyzed the P300 components and brain activity response induced by different complexity range images. We found that the brain topographic maps were different.

Fig 10 shows that the brain topographic maps varied between 400 ms to 600 ms in HIC, MIC, and LIC. We found a significant difference between HIC to MIC and LIC that the HIC peaked later than MIC and LIC (the HIC peaked at 475 ms, while the MIC and LIC peaked at 450 ms). This may mean that subjects need more time to identify the specific meaning of the image. In addition, between 525 ms and 600 ms, the brain activity of the LIC and MIC gradual decrease. However, in the HIC, the right frontal lobe has been active. This part is often associated with memory, semantics, images and other non-verbal ability. This result is interesting and reasonable. The subjects needed more time to analyze the specific meaning of complex images, in which case the right frontal lobe was active for a longer period of time. This result is interesting and it fits our expectations. However, due to the spatial distinguishability lack of EEG data, we cannot accurately determine the brain area of processes complex information. So combining fMRI or MEG with EEG data may be able to achieve better results, which will be the next step in the study.

thumbnail
Fig 10. The brain activity varied between 400 ms to 600 ms in different image complexity (HIC, MIC, and LIC).

https://doi.org/10.1371/journal.pone.0184713.g010

References

  1. 1. Thorpe S., Fize D., Marlot C., Speed of processing in the human visual system, Nature.1996; 381: 520–522. pmid:8632824
  2. 2. Wolpaw J.R., Birbaumer N., McFarland D.J., Pfurtscheller G., Vaughan T.M., Brain-computer interfaces for communication and control, Clin Neurophysiol. 2002; 113: 767–791. pmid:12048038
  3. 3. Wolpaw J.R., Mcfarland D.J., Control of a two-dimensional movement signal by a noninvasive brain-computer interface in humans, Proceedings of the National Academy of Sciences. 2005; 101: 17849–17854.
  4. 4. Chun M.M., Potter M.C., A two-stage model for multiple target detection in rapid serial visual presentation, Journal of Experimental psychology: Human perception and performance.1995; 21: 109–127. pmid:7707027
  5. 5. Polich J., Updating P300: an integrative theory of P3a and P3b, Clin Neurophysiol.2007; 118: 2128–2148. pmid:17573239
  6. 6. Yin E., Zeyl T., Saab R., Chau T., Hu D., Zhou Z., A Hybrid Brain-Computer Interface Based on the Fusion of P300 and SSVEP Scores, IEEE Transactions on Neural Systems & Rehabilitation Engineering A Publication of the IEEE Engineering in Medicine & Biology Society.2015; 23: 693–701.
  7. 7. Yin E., Zhou Z., Jiang J., Chen F., Liu Y., Hu D., A speedy hybrid BCI spelling approach combining P300 and SSVEP.2013; 61: 473–483.
  8. 8. Yin E., Zhou Z., Jiang J., Chen F., Liu Y., Hu D., A novel hybrid BCI speller based on the incorporation of SSVEP into the P300 paradigm, Journal of Neural Engineering.2013; 10: 026012. pmid:23429035
  9. 9. Yin E., Zeyl T., Saab R., Hu D., Zhou Z., Chau T., An Auditory-Tactile Visual Saccade-Independent P300 Brain-Computer Interface, International Journal of Neural Systems.2015; 26: 1650001. pmid:26678249
  10. 10. A.J. Ries, G.B. Larkin, Stimulus and Response-Locked P3 Activity in a Dynamic Rapid Serial Visual Presentation (RSVP) Task. DTIC Document.2013.
  11. 11. Touryan J., Gibson L., Horne J., Weber P., Real-time classification of neural signals corresponding to the detection of targets in video imagery. International Conference on Applied Human Factors and Ergonomics. 2010; p. 60.
  12. 12. Gonsalvez C.J., John P., P300 amplitude is determined by target-to-target interval, Psychophysiology.2002; 39: 388–396. pmid:12212658
  13. 13. Gerson A.D., Parra L.C., Sajda P., Cortically coupled computer vision for rapid image search, IEEE Trans Neural Syst Rehabil Eng.2005; 14: 174–179.
  14. 14. Parra L.C., Christoforou C., Gerson A.D., Dyrholm M., Luo A., Wagner M., et al. Spatiotemporal linear decoding of brain state, Signal Processing Magazine, IEEE.2008; 25: 107–115.
  15. 15. Sajda P., Pohlmeyer E., Wang J., Parra L.C., Christoforou C., Dmochowski J., et al. In a blink of an eye and a switch of a transistor: cortically coupled computer vision, Proceedings of the IEEE.2010; 98: 462–478.
  16. 16. Pohlmeyer E.A., Wang J., Jangraw D.C., Lou B., Chang S.F., Sajda P., Closing the loop in cortically-coupled computer vision: a brain-computer interface for searching image databases, J Neural Eng.2011;8: 036025. pmid:21562364
  17. 17. Alpert G.F., Manor R., Spanier A.B., Deouell L.Y., Geva A.B., Spatiotemporal Representations of Rapid Visual Target Detection: A Single-Trial EEG Classification Algorithm, IEEE Transactions on Biomedical Engineering.2014; 61: 2290–2303. pmid:24216627
  18. 18. Marathe A.R., Ries A.J., McDowell K., A novel method for single-trial classification in the face of temporal variability. Foundations of Augmented Cognition. 2013;2013: pp. 345–352.
  19. 19. Marathe A.R., Ries A.J., McDowell K., Sliding HDCA: single-trial EEG classification to overcome and quantify temporal variability, IEEE Trans Neural Syst Rehabil Eng.2014; 22: 201–211. pmid:24608681
  20. 20. Wang Y., Jiang L., Cai B., Wang Y., Zhang S., Zheng X., A closed-loop system for rapid face retrieval by combining EEG and computer vision. Neural Engineering (NER), 2015 7th International IEEE/EMBS Conference on IEEE.2015; pp. 130–133.
  21. 21. Cecotti H., Eckstein M.P., Giesbrecht B., Single-trial classification of event-related potentials in rapid serial visual presentation tasks using supervised spatial filtering, IEEE Transactions on Neural Networks & Learning Systems.2014; 25: 2030–2042.
  22. 22. Cecotti H., Marathe A., Ries A., Optimization of single-trial detection of event-related potentials through artificial trials, IEEE Transactions on Biomedical Engineering.2015; 62: 2170–2176. pmid:25823030
  23. 23. C. Zheng, C.P. Jose, O. Bing, D. Fraser, V. Anni, Marine animal classification using combined CNN and hand-designed image features, OCEANS'15 MTS/IEEE Washington. 2015:1–5.
  24. 24. Maxime O., Leon B., Ivan L., Josef S., Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks, IEEE Transactions on Computer Vision & Pattern Recognition. 2014: 1717–1724.
  25. 25. Griffin G., Holub A., Perona P., Caltech-256 Object Category Dataset, California Institute of Technology. 2007.
  26. 26. Sajda P., Gerson A., Parra L., High-throughput image search via single-trial event detection in a rapid serial visual presentation task. Neural Engineering, 2003. Conference Proceedings. First International IEEE EMBS Conference on.2003; pp. 7–10.
  27. 27. Krizhevsky A., Sutskever I., Hinton G.E., ImageNet Classification with Deep Convolutional Neural Networks, Advances in Neural Information Processing Systems.2012; 25: 1097–1105.
  28. 28. Jia Yangqing, Shelhamer Evan, Donahue Jeff, Karayev Sergey, Long Jonathan, Girshick Ross, et al. Caffe: Convolutional Architecture for Fast Feature Embedding, Proceedings of the 22nd ACM international conference on Multimedia. 2014: 675–678.
  29. 29. Cortes C., AUC Optimization vs. Error Rate Minimization, Advances in Neural Information Processing Systems.2004: 313–320.
  30. 30. Li F., Liu T., Wang F., Li H., Gong D., Zhang R., et al. Relationships between the resting-state network and the P3: Evidence from a scalp EEG study, Scientific Reports.2014; 5: 15129.
  31. 31. Agrawal P., Stansbury D., Malik J., Gallant J.L., Pixels to Voxels: Modeling Visual Representation in the Human Brain, Eprint Arxiv.2014: 1407.5104.
  32. 32. Güçlü U., van Gerven M.A., Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream, Journal of Neuroscience the Official Journal of the Society for Neuroscience.2015; 35: 10005–10014.
  33. 33. Cichy R.M., Khosla A., Pantazis D., Torralba A., Oliva A., Comparison of deep neural networks to spatio-temporal cortical dynamics of human visual object recognition reveals hierarchical correspondence, Scientific Reports.2015; 6: 27755.