Future technologies such as Brain-Computer Interaction Technologies (BCIT) or affective Brain Computer Interfaces (aBCI) will need to function in an environment with higher noise and complexity than seen in traditional laboratory settings, and while individuals perform concurrent tasks. In this paper, we describe preliminary results from an experiment in a complex virtual environment. For analysis, we classify between a subject hearing and reacting to an audio stimulus that is addressed to them, and the same subject hearing an irrelevant audio stimulus. We performed two offline classifications, one using BCILab [1], the other using LibSVM [2]. Distinct classifiers were trained for each individual in order to improve individual classifier performance [3]. The highest classification performance results were obtained using individual frequency bands as features and classifying with an SVM classifier with an RBF kernel, resulting in mean classification performance of 0.67, with individual classifier results ranging from 0.60 to 0.79. © 2011 Springer-Verlag.
CITATION STYLE
Lance, B., Gordon, S., Vettel, J., Johnson, T., Paul, V., Manteuffel, C., … Oie, K. (2011). Classifying high-noise EEG in complex environments for brain-computer interaction technologies. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6975 LNCS, pp. 467–476). https://doi.org/10.1007/978-3-642-24571-8_59
Mendeley helps you to discover research relevant for your work.