Sensors: Brain Computer Interfaces, A Review
Sensors: Brain Computer Interfaces, A Review
Sensors: Brain Computer Interfaces, A Review
3390/s120201211
OPEN ACCESS
sensors
ISSN 1424-8220
www.mdpi.com/journal/sensors
Review
Received: 29 December 2011; in revised form: 16 January 2012 / Accepted: 29 January 2012 /
Published: 31 January 2012
1. Introduction
A brain computer interface (BCI), also referred to as a brain machine interface (BMI), is a hardware
and software communications system that enables humans to interact with their surroundings,
without the involvement of peripheral nerves and muscles, by using control signals generated from
electroencephalographic activity. BCI creates a new non-muscular channel for relaying a person’s
intentions to external devices such as computers, speech synthesizers, assistive appliances, and neural
prostheses. That is particularly attractive for individuals with severe motor disabilities. Such an interface
would improve their quality of life and would, at the same time, reduce the cost of intensive care.
A BCI is an artificial intelligence system that can recognize a certain set of patterns in brain signals
following five consecutive stages: signal acquisition, preprocessing or signal enhancement, feature
extraction, classification, and the control interface [1]. The signal acquisition stage captures the brain
signals and may also perform noise reduction and artifact processing. The preprocessing stage
prepares the signals in a suitable form for further processing. The feature extraction stage identifies
discriminative information in the brain signals that have been recorded. Once measured, the signal is
mapped onto a vector containing effective and discriminant features from the observed signals. The
extraction of this interesting information is a very challenging task. Brain signals are mixed with other
signals coming from a finite set of brain activities that overlap in both time and space. Moreover, the
signal is not usually stationary and may also be distorted by artifacts such as electromyography (EMG)
or electrooculography (EOG). The feature vector must also be of a low dimension, in order to reduce
feature extraction stage complexity, but without relevant information loss. The classification stage
classifies the signals taking the feature vectors into account. The choice of good discriminative features
is therefore essential to achieve effective pattern recognition, in order to decipher the user’s intentions.
Finally the control interface stage translates the classified signals into meaningful commands for any
connected device, such as a wheelchair or a computer.
BCI technology has traditionally been unattractive for serious scientific investigation. The idea of
successfully deciphering thoughts or intentions by means of brain activity has often been rejected in
the past as very strange and remote. Hence investigation in the field of brain activity has usually been
limited to the analysis of neurological disorders in the clinic or to the exploration of brain functions in
the laboratory. The BCI design was considered too complex, because of the limited resolution and
reliability of information that was detectable in the brain and its high variability. Furthermore, BCI
systems require real-time signal processing, and up until recently the requisite technology either did
not exist or was extremely expensive [2].
However, this context has undergone radical change over the last two decades. BCI research, which
was confined to only three groups 20 years ago and only six to eight groups 10 years ago, is now a
flourishing field with more than 100 active research groups all over the World studying the topic [3].
The number of articles published regarding neural interface technology has increased exponentially
over the past decade [4]. Successful studies on brain signal phenomena have lent further weight to
these advances. The development of more and more inexpensive computer hardware and software has
allowed more sophisticated online analysis. Likewise, the chances of using BCIs as auxiliary
technology that might serve severely disabled people has increased social acceptance in this field and
the need to accelerate its progress. Interest in this technology is now found outside of the laboratory or
Sensors 2012, 12 1213
the clinic. Small specialized companies such as Emotiv [5] or Neurosky [6] have already developed
some initial applications oriented towards the general public. Nevertheless, despite these advances,
most BCI-based applications are still limited to the laboratory. Broader applicability of BCIs requires
greater ease of use, which in turn means reducing time spent on preparation, training and calibration [7].
BCI research is a relatively young multidisciplinary field integrating researchers from neuroscience,
physiology, psychology, engineering, computer science, rehabilitation, and other technical and
health-care disciplines. As a result, in spite of some notable advances, a common language has yet to
emerge, and existing BCI technologies vary, which makes their comparison difficult and, in
consequence, slows down the research. The community of BCI researchers has therefore stressed the
need to establish a general framework for BCI design [8]. Mason et al. [9], for example, proposed a
new functional model for BCI systems and taxonomy design.
This review of the state-of-the-art of BCI systems is arranged as follows: Section 2 discusses
existing neuroimaging approaches to BCIs and Section 3 describes the most commonly found control
signals in BCI systems. Section 4 briefly explains certain types of BCIs. Sections 5, 6 and 7,
respectively, cover the different signal processing methods used for feature extraction, artifact
reduction and feature classification. Section 8 provides an overview of BCI applications and, finally,
the conclusions are drawn in Section 9.
BCIs use brain signals to gather information on user intentions. To that effect, BCIs rely on a recording
stage that measures brain activity and translates the information into tractable electrical signals. Two
types of brain activities may be monitored: (i) electrophysiological and (ii) hemodynamic.
Electrophysiological activity is generated by electro-chemical transmitters exchanging information
between the neurons. The neurons generate ionic currents which flow within and across neuronal
assemblies. The large variety of current pathways can be simplified as a dipole conducting current
from a source to a sink through the dendritic trunk. These intracellular currents are known as primary
currents. Conservation of electric charges means that the primary currents are enclosed by extracellular
current flows, which are known as secondary currents [10]. Electrophysiological activity is measured
by electroencephalography, electrocorticography, magnetoencephalography, and electrical signal
acquisition in single neurons.
The hemodynamic response is a process in which the blood releases glucose to active neurons at a
greater rate than in the area of inactive neurons. The glucose and oxygen delivered through the blood
stream results in a surplus of oxyhemoglobin in the veins of the active area, and in a distinguishable
change of the local ratio of oxyhemoglobin to deoxyhemoglobin [11]. These changes can be quantified
by neuroimaging methods such as functional magnetic resonance and near infrared spectroscopy.
These kinds of methods are categorized as indirect, because they measure the hemodynamic response,
which, in contrast to electrophysiological activity, is not directly related to neuronal activity.
Most current BCIs obtain the relevant information from the brain activity through
electroencephalography. Electroencephalography is by far the most widely used neuroimaging
modality, owing to its high temporal resolution, relative low cost, high portability, and few risks to the users.
BCIs based on electroencephalography consist of a set of sensors that acquire electroencephalography signals
Sensors 2012, 12 1214
from different brain areas. However, the quality of electroencephalography signals is affected by scalp,
skull, and many other layers as well as background noise. Noise is key to electroencephalography and
to other neuroimaging methods, insofar as it reduces the SNR and therefore the ability to extract
meaningful information from the recorded signals.
Non-invasive approaches have successfully been used by severely and partially paralyzed patients
to reacquire basic forms of communication and to control neuroprostheses and wheelchairs [12–14].
Despite the outstanding utility of non-invasive approaches in BCI applications, motor recovery has
been limited, because of the need for brain signals with a higher resolution. Invasive recording
methods such as electrocorticography or intracortical neuron recording were introduced, in an effort to
improve the quality of brain signals monitored by BCIs. Most researchers agree that movement
restoration through prostheses with multiples degrees of freedom can only be achieved through
invasive approaches [15]. It is unlikely that the power of non-invasive modalities will be enhanced in
the near future. Accordingly, it would appear that invasive modalities are indispensable for accurate
neuroprostheses control. Nevertheless, this issue is not yet entirely clear and some opinions disagree
with this conjecture. Contrary to established opinion, Wolpaw [3] suggested that performance in
multidimensional control may be independent of the recording method. Further refinements of
recording and analysis techniques will probably increase the performance of both invasive and
non-invasive modalities. However, the latest studies in neuroprostheses control appear to indicate that
invasive modalities have inherent advantages in neuroprosthesis control applications [4].
Invasive modalities need to implant microelectrode arrays inside the skull that involves significant
health risks, which restricts their use to experimental settings. Two invasive modalities can be found in
BCI research: electrocorticography, which places electrodes on the surface of the cortex, either outside
the dura mater (epidural electrocorticography) or under the dura mater (subdural electrocorticography),
and intracortical neuron recording which implants electrodes inside the cortex. Several issues had to be
addressed, before they become suitable for long-term applications. First, tissue acceptance of the
microelectrode has to be addressed, for which reason proposals exist for electrodes with neurotropic
mediums that promote neuronal growth to improve biocompatibility [16]. Perhaps, the future of
nanotechnologies that might develop nano-detectors to be implanted inertly in the brain, may provide a
definite solution to the problems of long-term invasive applications. Second, a link between the
microelectrode and external hardware that uses wireless technology is needed to reduce the risks of
infection. Wireless transmission of neuronal signals has already been tested in animals [17]. And third,
continuous stress caused by plugging and unplugging the recording system may lead to tissue damage
or system failure.
Each neuroimaging modality is explained below. Firstly, electrophysiological methods such as
electroencephalography, electrocorticography, magnetoencephalography, and electrical signal acquisition
in single neurons will be discussed. Secondly, metabolic methods such as functional magnetic
resonance and near infrared spectroscopy will be described. Finally, functional imaging modalities are
listed in Table 1, along with information related to activity measured, temporal and spatial resolutions,
safety, and portability.
Sensors 2012, 12 1215
EEG measures electric brain activity caused by the flow of electric currents during synaptic
excitations of the dendrites in the neurons and is extremely sensitive to the effects of secondary
currents [10]. EEG signals are easily recorded in a non-invasive manner through electrodes placed on
the scalp, for which that reason it is by far the most widespread recording modality. However, it
provides very poor quality signals as the signals have to cross the scalp, skull, and many other layers.
This means that EEG signals in the electrodes are weak, hard to acquire and of poor quality. This
technique is moreover severely affected by background noise generated either inside the brain or
externally over the scalp.
The EEG recording system consists of electrodes, amplifiers, A/D converter, and a recording
device. The electrodes acquire the signal from the scalp, the amplifiers process the analog signal to
enlarge the amplitude of the EEG signals so that the A/D converter can digitalize the signal in a more
accurate way. Finally, the recording device, which may be a personal computer or similar, stores, and
displays the data.
The EEG signal is measured as the potential difference over time between signal or active electrode
and reference electrode. An extra third electrode, known as the ground electrode, is used to measure
the differential voltage between the active and the reference points. The minimal configuration for
EEG measurement therefore consists of one active, one reference, and one ground electrode.
Multi-channel configurations can comprise up to 128 or 256 active electrodes [18]. These electrodes
are usually made of silver chloride (AgCl) [19]. Electrode-scalp contact impedance should be between
1 kΩ and 10 kΩ to record an accurate signal [20]. The electrode-tissue interface is not only resistive
but also capacitive and it therefore behaves as a low pass filter. The impedance depends on several
factors such as the interface layer, electrode surface area, and temperature [20]. EEG gel creates a
conductive path between the skin and each electrode that reduces the impedance. Use of the gel is
cumbersome, however, as continued maintenance is required to assure a relatively good quality signal.
Electrodes that do not need to use of gels, called ‘dry’ electrodes, have been made with other materials
such as titanium and stainless-steel [21]. These kinds of electrodes may be ‘dry’ active electrodes,
which have preamplification circuits for dealing with very high electrode/skin interfacial
Sensors 2012, 12 1216
impedances [21,22], or ‘dry’ passive electrodes, which have no active circuits, but are linked to EEG
recording systems with ultra-high input impedance [23].
The amplitude of electrical bio-signals is in the order of microvolts. Consequently, the signal is very
sensitive to electronic noise. External sources such power-lines may generate background noise and
thermal, shot, flicker, and burst noises are generated by internal sources [24]. Design considerations
should be addressed to reduce the effects of the noise, such as electromagnetic interference shielding
or reduction for common mode signal, amongst others [20].
EEG comprises a set of signals which may be classified according to their frequency. Well-known
frequency ranges have been defined according to distribution over the scalp or biological significance.
These frequency bands are referred to as delta (δ), theta (θ), alpha (α), beta (β), and gamma (γ) from
low to high, respectively. Relevant characteristics of these bands are detailed below.
The delta band lies below 4 Hz, and the amplitude of delta signals detected in babies decreases as
they age. Delta rhythms are usually only observed in adults in deep sleep state and are unusual in
adults in an awake state. A large amount of delta activity in awake adults is abnormal and is related to
neurological diseases [25]. Due to low frequency, it is easy to confuse delta waves with artifact signals,
which are caused by the large muscles of the neck or jaw.
Theta waves lie within the 4 to 7 Hz range. In a normal awake adult, only a small amount of theta
frequencies can be recorded. A larger amount of theta frequencies can be seen in young children, older
children, and adults in drowsy, meditative or sleep states [25]. Like delta waves, a large amount of
theta activity in awake adults is related to neurological disease [25]. Theta band has been associated
with meditative concentration [26,27] and a wide range of cognitive processes such as mental
calculation [28], maze task demands [29], or conscious awareness [30].
Alpha rhythms are found over the occipital region in the brain [31]. These waves lie within the 8
to 12 Hz range. Their amplitude increases when the eyes close and the body relaxes and they attenuate
when the eyes open and mental effort is made [32]. These rhythms primarily reflect visual processing
in the occipital brain region and may also be related to the memory brain function [33]. There is also
evidence that alpha activity may be associated with mental effort. Increasing mental effort causes a
suppression of alpha activity, particularly from the frontal areas [34]. Consequently, these rhythms
might be useful signals to measure mental effort. Mu rhythms may be found in the same range as alpha
rhythms, although there are important physiological differences between both. In contrast to alpha
rhythms, mu rhythms are strongly connected to motor activities and, in some cases, appear to correlate
with beta rhythms [31,35].
Beta rhythms, within the 12 to 30 Hz range, are recorded in the frontal and central regions of the
brain and are associated with motor activities. Beta rhythms are desynchronized during real movement
or motor imagery [36]. Beta waves are characterized by their symmetrical distribution when there is no
motor activity. However, in case of active movement, the beta waves attenuate, and their symmetrical
distribution changes [36].
Gamma rhythms belong to the frequency range from 30 to 100 Hz. The presence of gamma waves
in the brain activity of a healthy adult is related to certain motor functions or perceptions, among
others [37]. Some experiments have revealed a relationship in normal humans between motor activities
and gamma waves during maximal muscle contraction [38]. This gamma band coherence is replaced
by a beta band coherence during weak contractions, suggesting a correlation between gamma or beta
Sensors 2012, 12 1217
cortical oscillatory activity and force [39]. Also, several studies have provided evidence for the role of
gamma activity in the perception of both visual and auditory stimuli [37,40–42]. Gamma rhythms are
less commonly used in EEG-based BCI systems, because artifacts such as electromyography (EMG) or
electrooculography (EOG) are likely to affect them [43]. Nevertheless, this range is attracting growing
attention in BCI research because, compared to traditional beta and alpha signals, gamma activity may
increase the information transfer rate and offer higher spatial specifity [44,45].
As explained above, EEG is recorded by electrodes. The electrodes placed over the scalp are
commonly based on the International 10–20 system [46], which has been standardized by the
American Electroencephalographic Society. The 10–20 system uses two reference points in the head to
define the electrode location. One of these reference points is the nasion, located at the top of the nose
at the same level as the eyes. The other reference point is the inion, which is found in the bony lump at
the base of the skull. The transverse and median planes divide the skull from these two points. The
electrode locations are determined by marking these planes at intervals of 10% and 20% (Figure 1).
The letters in each location corresponds to specific brain regions in such a way that A represents the
ear lobe, C the central region, Pg the nasopharyngeal, P the parietal, F the frontal, Fp the frontal polar,
and O the occipital area.
20
% Fp1 Fp2
%
Cz
% Fz
20 C3 Pz Fz
F3 F3 F4
F7 F8
Fp1 P3
20 %
20
F7
%
T3 A1 A2
10 % T5
C3 T3 C3 Cz C4 T4
Nasion A1
10 %
20 %
Pg1 T5 T6
P3 Pz P4
Inion
O1 O2
%
20
Inion 10 %
MEG is a non-invasive imaging technique that registers the brain’s magnetic activity by means of
magnetic induction. MEG measures the intracellular currents flowing through dendrites which produce
magnetic fields that are measurable outside of the head [47]. The neurophysiological processes that
produce MEG signals are identical to those that produce EEG signals. Nevertheless, while EEG is
extremely sensitive to secondary current sources, MEG is more sensitive to those of primary
Sensors 2012, 12 1218
currents [10]. The advantage of MEG is that magnetic fields are less distorted by the skull and scalp
than electric fields [48].
Magnetic fields are detected by superconducting quantum interferences devices, which are
extremely sensitive to magnetic disturbances produced by neural activity [49]. The electronic
equipment that measures magnetic brain activity is cooled to almost −273 degrees Celsius to facilitate
sensor superconductivity. MEG requires effective shielding from electromagnetic interferences. The
electronic equipment is installed inside a magnetically shielded room, which attenuates the effects of
magnetic fields from external sources.
MEG provides signals with higher spatiotemporal resolution than EEG, which reduces the training
time needed to control a BCI and speeds up reliable communications [50]. MEG has also been
successfully used to localize active regions inside the brain [51]. In spite of these advantageous
features, MEG is not often used in BCI design because MEG technology is too bulky and expensive to
become an acquisition modality suitable for everyday use. In 2005, Lal et al. [52] presented the first
online MEG-based BCI. Although further studies have followed [53–57], MEG-based BCIs, as
compared to EEG-based BCIs, are still at an early stage.
ECoG is a technique that measures electrical activity in the cerebral cortex by means of electrodes
placed directly on the surface of the brain. Compared to EEG, ECoG provides higher temporal and
spatial resolution as well as higher amplitudes and a lower vulnerability to artifacts such as blinks and
eye movement [58]. However, ECoG is an invasive recording modality which requires a craniotomy to
implant an electrode grid, entailing significant health hazards. For that reason, the first studies on
ECoG were with animals. Early studies involving animals evaluated the long-term stability of the
signals from the brain that ECoG could acquire [59–62]. The results showed that subdural electrodes
could provide stable signals over several months. Nevertheless, the long-term stability of the signals
acquired by ECoG is currently unclear. More recent experiments with monkeys have shown that ECoG
can perform at a high level for months without any drift in accuracy or recalibration [63]. The hand
positions and arm joint angles could be successfully decoded during asynchronous movements. These
studies have also developed minimally invasive protocols to implant the ECoG probes [64].
In humans, ECoG has been used for the analysis of alpha and beta waves [65] or gamma waves [66,67]
produced during voluntary motor action. With regard to the use of ECoG in BCIs systems,
Levine et al. [68] designed a BCI which classified motor actions on the basis of the identification of
the event-related potentials (ERP) using ECoG. Leuthardt et al. [69] showed for the first time that an
ECoG-based BCI could provide information to control a one-dimensional cursor, as this information is
more precise and more quickly acquired than by EEG-based BCIs. Some years later, Schalk et al. [70]
presented a more advanced ECoG-based BCI which allowed the user to control a two-dimensional
cursor. The results of all these studies might make it more feasible for people with severe motor
disabilities to use ECoG-based BCIs for their communication and control needs.
Sensors 2012, 12 1219
Intracortical neuron recording is a neuroimaging technique that measures electrical activity inside
the gray matter of the brain. It is an invasive recording modality that needs to implant microelectrode
arrays inside the cortex to capture spike signals and local field potentials from neurons.
Three signals can be obtained by intracortical neuron recording: single-unit activity (SUA),
multi-unit activity (MUA), and local field potentials (LFPs) [47]. SUA is obtained by high-pass
filtering (>300 Hz) of the signal of a single neuron. MUA is obtained in the same way, but the signals
may come from multiple neurons. LFPs are extracted by low-pass filtering (<300 Hz) of the neuron
activity in the vicinity of an electrode tip. LFPs are analog signals whereas SUA and MUA measure
the spiking activity of single neurons and can be reduced to discrete events in time [47].
Intracortical neuron recording provides much higher spatial and temporal resolution than EEG
recording. Hence the intracortical signals may be easier to use than EEG signals. However, signal quality
may be affected by the reaction of cerebral tissue to the implanted recording microelectrode [71] and
by changes in the sensitivity of the microelectrode, which may be progressively damaged over the
course of days and years [72]. The user can naturally adapt to these slow changes in the relative
sensitivity of the microelectrode, without the need for specific retraining. Nevertheless, periodic
recalibrations of electrode sensitivity may be necessary [73].
The first attempts in the intracortical neuron recording field were made in animals. Multielectrode
arrays have been used to record neural activity from the motor cortex in monkeys or rats during
learned movements [74–76]. These initial studies have shown that intracortical neuron recordings can
indicate the nature of a movement and its direction. These studies do not reveal whether the same
patterns will be present when the real movements are not made. In that regard, Taylor and Schwartz [77]
experimented with rhesus macaques, which made real and virtual arm movements in a computer. The
results suggested that the same patterns persisted. The most recent studies with monkeys investigated
the control of prosthetic devices for direct real-time interaction with the physical environment [78–81].
With regard to the application of intracortical neuron recording in BCI systems, microelectrode
arrays such as the Utah Intracortical Electrode Array (UIEA) have been reported as a suitable means of
providing simultaneous and proportional control of a large number of external devices [72]. Also,
Kennedy et al. [82] employed cortical control signals to design a BCI that allowed users to control
cursor movement and flexion of a cyber-digit finger on a virtual hand.
fMRI is a non-invasive neuroimaging technique which detects changes in local cerebral blood
volume, cerebral blood flow and oxygenation levels during neural activation by means of
electromagnetic fields. fMRI is generally performed using MRI scanners which apply electromagnetic
fields of strength in the order of 3T or 7T. The main advantage of the use of fMRI is high space
resolution. For that reason, fMRI have been applied for localizing active regions inside the brain [83].
However, fMRI has a low temporal resolution of about 1 or 2 seconds. Additionally, the hemodynamic
response introduces a physiological delay from 3 to 6 seconds [84]. fMRI appears unsuitable for rapid
communication in BCI systems and is highly susceptible to head motion artifacts.
Sensors 2012, 12 1220
In BCI systems, fMRI is typically used to measure the Blood Oxygen Level Dependent (BOLD)
during neuronal activation [85]. Although the BOLD signal is not directly related to neuronal activity,
a correspondence between both does exist [86]. The use of fMRI in BCI technology is relatively
recent. Before the emergence of real-time fMRI, brain activity recording by fMRI has traditionally
taken a long time. The data acquired by fMRI techniques were processed offline and the results only
became available after several hours or even days [87]. fMRI-based BCIs have been made possible,
thanks to the development of real-time fMRI [84,88,89]. The information transfer rate in fMRI-based
BCIs is between 0.60 and 1.20 bits/min [90]. Non-clinical fMRI applications are not expected because
fMRI requires overly bulky and expensive hardware.
NIRS is an optical spectroscopy method that employs infrared light to characterize noninvasively
acquired fluctuations in cerebral metabolism during neural activity. Infrared light penetrates the skull
to a depth of approximately 1–3 cm below its surface, where the intensity of the attenuated light allows
alterations in oxyhemoglobin and deoxyhemoglobin concentrations to be measured. Due to shallow
light penetration in the brain, this optical neuroimaging technique is limited to the outer cortical layer.
In a similar way to fMRI, one of the major limitations of NIRS is the nature of the hemodynamic
response, because vascular changes occur a certain number of seconds after its associated neural
activity [91]. The spatial resolution of NIRS is quite low, in the order of 1 cm [92]. Nevertheless,
NIRS offers low cost, high portability, and an acceptable temporal resolution in the order of
100 milliseconds [93].
A NIRS system consists of a light source, a driving electronic device, a light detector, signal
processing devices, and a recording device. The light source is an infrared emitting diode (IRED)
placed in direct contact with the scalp. The driving electronic device is an electronic circuit that
controls the IRED in order to modulate the light. The light detector is a photodiode placed right next to
the light source. The signal processing devices are amplifiers and filters that process the electrical
signal and reduce the noise due to ambient light. The recording device is a personal computer or any
other device that digitalizes, stores, and displays the electrical signal.
Ensuring good coupling light from the optical sources and detectors to and from the subject’s head
is not a trivial issue. Head motions or hair obstruction can worsen performance and signal quality [91].
Good quality signals and noise reduction, especially background noise induced by head motions, are
important requirements in real time BCI systems. Hair obstruction can be overcome by combing the
hair out of the photons’ path by means of hair gel and hair clips [91]. Noise can be reduced partially by
bandpass filtering, moving averaging, and Wiener filtering. These classes of algorithms usually fail to
remove abrupt spike-like noise produced by head motion [94]. Head motion artifacts can be minimized
by ensuring rigid optode positioning. Solutions have been introduced that are based on helmets,
thermoplastic molded to the contours of each subject’s head, spring-loaded fibers attached to
semi-rigid plastic forms, and fibers embedded in neoprene rubber forms [95]. Background noise effects
can also be attenuated by exploiting the strong statistical association between oxygenated and
deoxygenated hemoglobin dynamics [96].
Sensors 2012, 12 1221
The purpose of a BCI is to interpret user intentions by means of monitoring cerebral activity. Brain
signals involve numerous simultaneous phenomena related to cognitive tasks. Most of them are still
incomprehensible and their origins are unknown. However, the physiological phenomena of some
brain signals have been decoded in such way that people may learn to modulate them at will, to enable
the BCI systems to interpret their intentions. These signals are regarded as possible control signals
in BCIs.
Numerous studies have described a vast group of brain signals that might serve as control signals in
BCI systems. Nevertheless, only those control signals employed in current BCI systems will be
discussed below: visual evoked potentials, slow cortical potentials, P300 evoked potentials, and
sensorimotor rhythms. All the signal controls are listed in Table 2, along with some of their main features.
Number of Information
Signal Physiological phenomena Training
choices transfer rate
60–100
VEP Brain signal modulations in the visual cortex High No
bits/min
Low (2 or 4, very
SCP Slow voltages shift in the brain signals Yes 5–12 bits/min
difficult)
20–25
P300 Positive peaks due to infrequent stimulus High No
bits/min
Sensorimotor Modulations in sensorimotor rhythms
Low (2, 3, 4, 5) Yes 3–35 bits/min
rhythms synchronized to motor activities
VEPs are brain activity modulations that occur in the visual cortex after receiving a visual
stimulus [101]. These modulations are relatively easy to detect since the amplitude of VEPs increases
enormously as the stimulus is moved closer to the central visual field [102].
VEPs may be classified according to three different criteria [103]: (i) by the morphology of the
optical stimuli, (ii) by the frequency of visual stimulation; and (iii) by field stimulation. According to
the first criterion, VEPs may be caused by using flash stimulation or using graphic patterns such as
checkerboard lattice, gate, and random-dot map. According to the frequency, VEPs can also be
classified as transient VEPs (TVEPs) and as steady-state VEPs (SSVEPs). TVEPs occur when the
Sensors 2012, 12 1222
frequency of visual stimulation is below 6 Hz, while SSVEPs occur in reaction to stimuli of a higher
frequency [101,104]. Lastly, according to the third criterion, VEPs can be divided into whole field
VEPs, half field VEPs, and part field VEPs depending on the area of on-screen stimulus. For instance,
if only half of the screen displays graphics, the other half will not display any visual stimulation, and
the person will look at the centre of the screen, which will induce a half field VEP.
TVEPs can be elicited by any change in the visual field. Those used most frequently are TVEPs are:
(i) flash TVEPs that are caused by flashing lights; (ii) pattern onset/offset TVEPs that are caused by
letting a pattern appear abruptly on a diffuse background; and (iii) pattern reversal TVEPs that are
caused by reversing the phase of a pattern i.e., a checkerboard lattice that changes the checks from
black to white and from white to black abruptly [105].
The evoked responses vary with the stimulus presented. Flash TVEPs present a series of negative
and positive peaks. The most prominently peaks are negative (N2) and positive (P2) peaks at around
90 ms and 120 ms respectively [105]. Pattern onset/offset TVEPs have three main peaks: C1 (positive,
75 ms), C2 (negative, 125 ms), and C3 (positive, 150 ms) [105]. Pattern reversal TVEPs usually present
one negative peak at 75 ms, one positive peak at 100 ms, and one negative peak at 135 ms [105].
SSVEPs are elicited by the same visual stimulus. In this case, the stimulus changes at a frequency
higher than 6 Hz. If the stimulus is a flash, SSVEP shows a sinusoidal-like waveform, the fundamental
frequency of which is the same as the blinking frequency of the stimulus. If the stimulus is a pattern,
the SSVEP occurs at the reversal rate and at their harmonics [106]. In contrast to TVEP, constituent
discrete frequency components of SSVEPs remain closely constant in amplitude and phase over long
periods of time [107]. SSVEPs are less susceptible than TVEPs to artifacts produced by blinks and eye
movements [108] and to electromyographic noise contamination [109]. Indeed, TVEPs not are
typically used for BCI.
SSVEP-based BCIs allow users to select a target by means of an eye-gaze. The user visually fixes
attention on a target and the BCI identifies the target through SSVEP features analysis. Considering a
BCI as a communications channel, SSVEP-based BCIs can be classified into three categories
depending on the specific stimulus sequence modulation in use [110]: time modulated VEP (t-VEP)
BCIs, frequency modulated VEP (f-VEP) BCIs, and pseudorandom code modulated VEP (c-VEP)
BCIs. VEPs that react to different stimulus sequences should be orthogonal or near orthogonal to each
other in some domain to ensure reliable identification of the target [110]. In a t-VEP BCI, the flash
sequences of different targets are orthogonal in time. That is, the flash sequences for different targets
are either strictly non-overlapping or stochastic. In an f-VEP BCI, each target is flashed at a unique
frequency, generating a periodic sequence of evoked responses with the same fundamental frequency
as its harmonics. In a c-VEP BCI, pseudo-random sequences are used. The duration of ON and OFF
states of each target’s flash is determined by a pseudorandom sequence. Signal modulations can
optimize the information transfer rate. Indeed, code modulation provides the highest communication
speed. Table 3 summarizes the features of each modulation.
Sensors 2012, 12 1223
The typical VEP-based BCI application displays flashing stimuli, such as digits or letters, on a
screen to induce SSVEPs while the user stares at one of the symbols. The user can move their gaze to
the flashing digits or letters, in order to communicate with the computer [111]. The advantage of this
type of control signal is that very little training is required. However, it presents the drawback that the
user has to watch the screen and keep his eyes fixed on one point. This type of control signal can only
be used for exogenous BCIs (see Section 0). Therefore, VEPs are not suitable for patients in advanced
stages of Amyotrophic Lateral Sclerosis (ALS) or with uncontrollable eye or neck movements. Some
independent SSVEP-based BCIs that are controlled by the attention of the user have been introduced to
overcome this drawback [112,113].
SSVEP are usually elicited through light-emitting diodes (LEDs), cathode-ray tube (CRT) monitors,
or liquid crystal display (LCD). LEDs outperform LCD or CRT stimulators but they need more
complex hardware. LCD and CRT monitors make the target presentation easier than LED stimulators,
because both systems can easily be connected to a PC. However, LED stimulators may be preferable
for a multiple target BCI, because the refresh rate of an LCD or CRT monitor can limit the number of
targets. LED stimulators offer more versatility because the flickering frequency and phase of each
LED can be controlled independently by a programmable logic device [114]. The stimulation decision
can be made on the basis of the number of choices that the BCI offers [115]. LCD screens are optimal
for low complexity BCI (less than 10 choices), because they induce less eye-tiredness than CRT
screens. For medium complexity BCI (10–20 choices), LCD or CRT screens are optimal. For high
complexity BCI (more than 20 commands), LED are preferred.
SCPs are slow voltage shifts in the EEG that last a second to several seconds. SCPs belong to the
part of the EEG signals below 1 Hz [116]. SCPs are associated with changes in the level of cortical
activity. Negative SCPs correlate with increased neuronal activity, whereas positive SCPs coincide
with decreased activity in individual cells [116]. These brain signals can be self-regulated by both
healthy users and paralyzed patients to control external devices by means of a BCI. SCP shifts can be
used to move a cursor and select the targets presented on a computer screen [117].
Sensors 2012, 12 1224
People can be trained to generate voluntary SCP changes using a thought-translation device [117].
The thought-translation device is a tool used for self-regulation SCP training, which shows
visual-auditory marks so that the user can learn to shift the SCP. The thought-translation device
typically comprises a cursor on a screen in such a way that the vertical position of the cursor constantly
reflects the amplitude of SCP shifts. Although most thought-translation devices show continuous
feedback, it is possible to train SCP self-modulation in the absence of continuous feedback [118].
Success in SCP self-regulation training depends on numerous factors, such as the patient’s
psychological and physical state, motivation, social context, or the trainer-patient relationship [117]. It
is known that the learning capability of the user drastically affects SCP modulation training.
Self-regulation training is therefore strongly recommended for patients at the early stage of a
progressive disease [117]. Furthermore, initial SCP modulation skills have an effect on future
performance following training [119]. Therefore, the value of SCPs as a suitable control signal for each
patient can only be determined on the basis of initial trials. Other factors, such as sleep quality, pain,
and mood also have an influence on self-regulation performance [117]. Their effects are not identical
for all patients and further investigation is certainly needed to establish general rules on this matter.
Self-regulation of SCPs has been tested extensively with patients suffering from ALS [120–122].
Typical accuracy rates achieved for SCP classification are acceptable and vary between 70 and 80 per
cent, but the rates of information provided by SCP-based BCI are relatively low. Besides, longer
training is required to use SCP-based BCI and it is likely that users will need continuous practice for
several months.
P300 evoked potentials are positive peaks in the EEG due to infrequent auditory, visual, or
somatosensory stimuli. These endogenic P300 responses are elicited about 300 ms after attending to an
oddball stimulus among several frequent stimuli [123,124]. Some studies have proven that the less
probable the stimulus, the larger the amplitude of the response peak [125]. The use of P300-based
BCIs does not require training. However, the performance may be reduced because the user gets used
to the infrequent stimulus and consequently P300 amplitude is decreased [126].
A typical application of a BCI based on visual P300 evoked potentials comprises a matrix of letters,
numbers, or other symbols or commands [123,127,128]. The rows or columns of this matrix are
flashed at random while the EEG is monitored. The user gazes at the desired symbol and counts how
many times the row or column containing the desired choice flashes. P300 is elicited only when the
desired row or column flashes. Thus, the BCI uses this effect to determine the target symbol. Due to
the low signal-to-noise ratio in EEG signals, the detection of target symbols from a single trial is very
difficult. The rows or columns must be flashed several times for each choice. The epochs corresponding to
each row or column are averaged over the trials, in order to improve their accuracy. However, these
repetitions decrease the number of choices per minute, e.g., with 15 repetitions, only two characters are
spelled per minute [123]. Although most of the applications based on P300 evoked potentials employ
visual stimuli, auditory stimuli have been used for people with visual impairment [129].
P300-based BCIs provide a very low rate of information transmission because the classifier based
on an average is too simple, and the accuracy of P300 potential detection is too low [130].
Sensors 2012, 12 1225
Consequently, too many trials are required to select a single symbol in the matrix. Accuracy of
P300-based BCIs can be improved, while using a more complicated classifier than a simple average to
ensure that the number of repetitions remain unaffected [130,131]. Other studies have proven that the
detection accuracy of visual P300 evoked potentials also depends on the properties of the visual matrix
such as the dimensions or colors of the symbols. Performance decreases when matrices with smaller
symbols are used [132], and it is enhanced when a green and blue chromatic flicker matrix is used,
rather than a gray and black one [133,134].
Information transmission rates provided by P300-based BCI can be also improved by considering
the BCI as a noisy transmission system. BCI can therefore benefit from the use of error correcting
codes [135]. However, optimizing the code solely according to the maximal minimum-Hamming-distance
implies an increase in target frequency of target stimuli which might violate physiological constraints
leading to difficulties in classifying the individual ERPs, due to overlap and refractory effects. Further,
overlap and refractory effects are generally the main error source in these kinds of BCIs [136]. Some
recent novel approaches have tried to reduce them, by superimposing the targets on a checkerboard [137]
or by using alternative stimulus type methods based on motion [136].
The P300 response is not markedly affected by whether or not the subject gazes directly at the
target, in contrast to the VEP response, which is larger when the target is foveated [138]. This
distinction is important for clinical applications, because eye movements are often impaired or lost in
the target population. Nevertheless, the performance of a P300-based BCI is substantially improved
when subjects gaze at the desired item [138]. Therefore, the performance of the visual P300-based
BCIs depends not only on the P300-evoked potential, but also on the VEP response that, in turn,
strongly depends on eye-gaze direction.
Sensorimotor rhythms comprise mu and beta rhythms, which are oscillations in the brain activity
localized in the mu band (7–13 Hz), also known as the Rolandic band, and beta band (13–30 Hz),
respectively. Both rhythms are associated in such a way that some beta rhythms are harmonic mu
rhythms, although some beta rhythms may also be independent [139]. The amplitude of the
sensorimotor rhythms varies when cerebral activity is related to any motor task although actual
movement is not required to modulate the amplitude of sensorimotor rhythms [140,141]. Similar
modulation patterns in the motor rhythms are produced as a result of mental rehearsal of a motor act
without any overt motor output [140]. Sensorimotor rhythms have been used to control BCIs, because
people can learn to generate these modulations voluntarily in the sensorimotor rhythms [36,142].
Sensorimotor rhythms can endure two kinds of amplitude modulations known as event-related
desynchronization (ERD) and event-related synchronization (ERS) that are generated sensory
stimulation, motor behavior, and mental imagery [36]. ERD involves an amplitude suppression of the
rhythm and ERS implies amplitude enhancement. Figure 2 (left panel) shows the temporal behavior of
ERD and ERS during a voluntary movement experiment which involves brisk finger lifting [36]. The
mu band ERD starts 2.5 s before movement on-set, reaches the maximal ERD shortly after
movement-onset, and recovers its original level within a few seconds. In contrast, the beta rhythm
shows a short ERD during the movement initiation of movement, followed by ERS that reaches the
Sensors 2012, 12 1226
maximum after movement execution. This ERS occurs while the mu rhythm is still attenuated.
Figure 2 also shows the gamma oscillation (36–40 Hz), which is another rhythm related to motor tasks
as well [36]. Gamma rhythms reveal an ERS shortly before movement-onset. Finally, the right panel of
Figure 2 illustrates that simultaneous ERD and ERS are possible at different scalp locations [36].
Figure 2. Left panel: Superimposed band power time courses computed for three different
frequency bands (10–12 Hz, 14–18 Hz, and 36–40 Hz) from EEG trials recorded from
electrode position C3 during right index finger lifting. EEG data triggered with respect to
movement-offset (vertical line at t = 0 s); Right panel: Examples of ongoing EEG
recorded during right finger movement (adapted from [36]).
250
Post-movement ERD
200 C3 beta ERS Fz
Gamma ERS 14-18 Hz
ERD / ERS %
150 36-40 Hz C3
C4 ERS
100
Mu ERD Cz
50 10-12 Hz
P3
0 P4
-50 Oz
100
-6 -5 -4 -3 -2 -1 0 1 2 3 -5 -4 -3 -2 -1 0 1 2 3 s
Sensorimotor rhythms are related to motor imagery without any actual movement [141]. This makes
it possible to use sensorimotor rhythms for the design of endogenous BCIs, which are more useful than
exogenous BCIs. Nevertheless, self-control of sensorimotor rhythms is not easy, and most people have
difficulties with motor imagery. People tend to imagine visual images of related real movements,
which is not sufficiently useful for a BCI system, because the patterns of these sensorimotor rhythms
differ from actual motor imagery. User training should emphasize kinesthetic experiences instead of
visual representations of actions [143]. Motor imagery training is traditionally based on visual or
auditory feedback [144]. This kind of training asks the users to perform a certain motor imagery task,
and then the sensorimotor rhythms are extracted and classified by comparing them with a reference.
Finally, visual or auditory feedback is provided to the participant according to the success of the result.
This kind of training has been widely used although usually its effectiveness was not very high [145].
Hwang et al. [145] presented more effective motor imagery training based on a system that displayed
real-time cortical activity as feedback, which allowed the users to watch their own cortical activity
through a real-time monitoring system.
Sensorimotor rhythms have been investigated extensively in BCI research. Well-known BCI
systems such as Wadsworth [146], Berlin [147], or Graz [148] BCIs employ sensorimotor rhythms as
control signals. The BCIs based on sensorimotor rhythms can operate in either synchronous or
asynchronous mode. The latest advances in the field of BCIs based on sensorimotor rhythms have
shown that it is possible to predict human voluntary movements before they occur based on the
modulations in sensorimotor rhythms [149]. Furthermore, this prediction could be provided without
the user making any movements at all.
Sensors 2012, 12 1227
4. Types of BCIs
The BCIs can be categorized into (i) exogenous or endogenous and (ii) synchronous (cue-paced) or
asynchronous (self-paced). Types of BCI are listed in Tables 4 and 5, along with information related to
brain signals that can be modulated to convey information as well as advantages and disadvantages.
Also, BCIs can be classified into dependent and independent [2]. This distinction will not be detailed
in this review because it is very similar to exogenous and endogenous distinction. Advantages and
disadvantages in both taxonomies are analogous.
According to the nature of the signals used as input, BCI systems can be classified as either
exogenous or endogenous. Exogenous BCI uses the neuron activity elicited in the brain by an external
stimulus such as VEPs or auditory evoked potentials [150]. Exogenous systems do not require
extensive training since their control signals, SSVEPs and P300, can be easily and quickly set-up.
Besides, the signal controls can be realized with only one EEG channel and can achieve a high
information transfer rate of up to 60 bits/min. On the other hand, endogenous BCI is based on
self-regulation of brain rhythms and potentials without external stimuli [150]. Through neurofeedback
training, the users learn to generate specific brain patterns which may be decoded by the BCI such as
modulations in the sensorimotor rhythms [151] or the SCPs [117]. The advantage of an endogenous
BCI is that the user can operate the BCI at free will and move a cursor to any point in a
Sensors 2012, 12 1228
two-dimensional space, while an exogenous BCI may constrain the user to the choices presented. Also,
endogenous BCI are especially useful for users with advanced stages of ALS or whose sensory organs
are affected. Table 4 summarizes the differences between exogenous and endogenous BCIs.
According to the input data processing modality, BCI systems can be classified as synchronous or
asynchronous. Synchronous BCIs analyze brain signals during predefined time windows. Any brain
signal outside the predefined window is ignored. Therefore, the user is only allowed to send commands
during specific periods determined by the BCI system. For example, the standard Graz BCI [148]
represents a synchronous BCI system. The advantage of a synchronous BCI system is that the onset of
mental activity is known in advance and associated with a specific cue [152]. Moreover, the patients
may also perform blinks and other eye movements, which would generate artifacts, if the BCI did not
analyze the brain signals to avoid their misleading effects. This simplifies the design and evaluation of
synchronous BCI. Asynchronous BCIs continuously analyze brain signals no matter when the user
acts. They offer a more natural mode of human-machine interaction than synchronous BCI. However,
asynchronous BCIs are more computation demanding and complex. Table 5 summarizes the
differences between synchronous and asynchronous BCIs.
Different thinking activities result in different patterns of brain signals. BCI is seen as a pattern
recognition system that classifies each pattern into a class according to its features. BCI extracts some
features from brain signals that reflect similarities to a certain class as well as differences from the rest
of the classes. The features are measured or derived from the properties of the signals which contain
the discriminative information needed to distinguish their different types.
The design of a suitable set of features is a challenging issue. The information of interest in brain
signals is hidden in a highly noisy environment, and brain signals comprise a large number of
simultaneous sources. A signal that may be of interest could be overlapped in time and space by
multiple signals from different brain tasks. For that reason, in many cases, it is not enough to use
simple methods such as a band pass filter to extract the desired band power.
Brain signals can be measured through multiples channels. Not all information provided by the
measured channels is generally relevant for understanding the underlying phenomena of interest.
Dimension reduction techniques such as principal component analysis or independent component
analysis can be applied to reduce the dimension of the original data, removing the irrelevant and
redundant information. Computational costs are thereby reduced.
Brain signals are inherently non-stationary. Time information about when a certain feature occurs
should be obtained. Some approaches divide the signals into short segments and the parameters can be
estimated from each segment. However, the segment length affects the accuracy of estimated features.
FFT performs very poorly with short data segments [153]. Wavelet transform or adaptive autoregressive
components are preferred to reveal the non-stationary time variations of brain signals. Also, a novel
technique called stationary subspace analysis (SSA) has recently been introduced to deal with the
non-stationarity of EEG signals [154]. SSA decomposes multivariate time series into stationary and
non- stationary components.
Sensors 2012, 12 1229
Multiples features can be extracted from several channels and from several time segments before
being concatenated into a single feature vector. One of the major difficulties in BCI design is choosing
relevant features from the vast number of possible features. High dimensional feature vectors are not
desirable due to the “curse of dimensionality” in training classification algorithms (see next section).
The feature selection may be attempted examining all possible subsets of the features. However, the
number of possibilities grows exponentially, making an exhaustive search impractical for even a
moderate number of features. Some more efficient optimization algorithms can be applied with the aim
of minimizing the number of features while maximizing the classification performance.
This section discusses methods to obtain the relevant characteristics of brain signals as well as
feature selection methods. Firstly, dimensional reduction methods, such as principal component
analysis or independent component analysis are explained. Secondly, time and/or frequency methods,
such as matched filtering or wavelet transform, and parametric modeling, such as autoregressive
component, are also surveyed. Thirdly, an explanation is given of the common spatial pattern
algorithm. This method designs a preprocessing spatial filter, by means of spatial covariance from input
data and signal whitening, that enhances the difference between classes before the feature extraction
stage. And, finally, feature selection methods such as genetic algorithms or sequential selection are
included. All these methods, including feature extraction and feature selection methods, are listed
respectively in Tables 6 and 7 along with information on their properties and BCI applications.
Table 6. Cont.
− Spectrum model
− High frequency resolution for short time segments
AR [170,172]
− Not suitable for non-stationary signals
− Adaptive version of AR: MVAAR
− Detects a specific pattern on the basis of its matches with
predetermined known signals or templates
Time- MF [151,173]
− Suitable for detection of waveforms with consistent temporal
frequency
characteristics
− Provides both frequency and temporal information
CWT [179,180]
− Suitable for non-stationary signals
− Provides both frequency and temporal information
DWT − Suitable for non-stationary signals [181,182]
− Reduces the redundancy and complexity of CWT
PCA is a statistical features extraction method that uses a linear transformation to convert a set of
observations possibly correlated into a set of uncorrelated variables called principal components.
Linear transformation generates a set of components from the input data, sorted according to their
variance in such a way that the first principal component has the highest possible variance. This
variance allows PCA to separate the brain signal into different components.
PCA projects the input data on a k-dimension eigenspace of k eigenvectors, which are calculated
from the covariance matrix ∑ of the training data p = [p1 p2 … pn] [155]. pi is i-th d-dimension training
sample, and n is the number of samples.
The covariance matrix ∑ is computed as:
∑ (1)
eigenvalue represent the principal components of the training dataset p. PCA selects that k, with k < d,
eigenvectors having the largest eigenvalues. These selected eigenvectors serve to build a projection
matrix A that will be used to extract the feature vector from the test data q. The k eigenvectors are
sorted into columns in Matrix A, such that the first column of A corresponds to the largest eigenvalue.
Finally, PCA computes the feature vector v from the data in matrix A, by projecting the test data q onto
the new subspace, such that:
(2)
where, ∑ represents the mean vector of training samples pi.
PCA is also a procedure to reduce the dimension of the feature. Since the number of columns is less
than the number of eigenvectors, the dimension of the output projected data is less than the dimension
of the input data. This decrease in dimensionality can reduce the complexity of the subsequent
classifying step in a BCI system.
PCA does not always guarantee a good classification since the best discriminating components may
not figure among the largest principal components [156]. PCA reduces data dimension by seeking a
new optimal representation of data in terms of minimal mean-square-error between the representation
and the original data. It will not guarantee that the discriminative features are optimal for classification.
Despite this shortcoming, it has been proven that PCA is a reliable noise reduction method.
With regard to the applications of PCA in BCI systems, PCA has been used to identify the
artifactual components in a reasonably successful way in EEG signals and to reconstruct the signals
without the artifactual components [157,158]. Nevertheless, the artifacts must not be correlated with
the EEG signal for PCA to function in this way. PCA has also been employed, in order to reduce
feature space dimensionality [155].
ICA is a statistical procedure that splits a set of mixed signals into its sources with no previous
information on the nature of the signal. The only assumption involved in ICA is that the unknown
underlying sources are mutually independent in statistical terms. ICA assumes that the observed EEG
signal is a mixture of several independent source signals coming from multiple cognitive activities or
artifacts. ICA therefore expresses the resulting EEG signal x(t) in relation to their sources s(t) as:
(3)
where, f is any unknown mixer function, and n(t) is an additive random noisy vector. The dimension of
the input vector s(t) depends on the number of sources. The dimension of output vector x(t) is equal to
the number of measured data channels. The number of sources is usually assumed to be less than or
equal to the number of channels although more generalized ICA methods are possible [159].
The whole ICA problem consists in the calculation of the unmixing function by inverting f and
obtaining an estimation of s(t), by mapping x(t) to the source space. To solve the problem, ICA can fall
into two different models on the basis of f, which may be either a linear or nonlinear function. The
nonlinear assumption is suitable in those cases where the linear model might be too simple to describe
the observed data x(t). However, the nonlinear problem is usually too complex and generally
intractable due to its high number of indeterminations. The assumption of a linear mixing function
Sensors 2012, 12 1232
simplifies Equation (3). It is possible to rewrite it as a matrix multiplication where A is the mixing
matrix. The Equation (4) gives the mathematical expression of the linear ICA model:
(4)
Although the approximation given by Equation (4) can be considered too simple, it works
reasonably well in brain signal processing applications. Furthermore, it is possible to remove the noise
term n(t) from Equation (4), by assuming that the observed data is noiseless or that the noise is too
weak for consideration [160,161]. Finally, s(t) and A are obtained from x(t) by means of certain
algorithms, such as Infomax [162] or further modification of the Infomax [163].
ICA has traditionally been used as a preprocessing tool before the feature extraction step, in order to
remove ocular artifacts in BCI systems [164–166]. Although ICA has been proven to be a powerful
and robust tool for artifact removal in signal analysis, some studies have indicated that artifact
suppression may also corrupt the power spectrum of the underlying neural activity [167]. In addition,
ICA requires that the artifacts are independent in relation to the EEG signal.
It is also possible to find authors that have employed ICA as a classifier. ICA can be modified to
classify EEG signals by fitting the generative ICA model to each task and employing Bayes’ rule to
create the classifier [168].
AR spectral estimation is a method for modeling signals. AR models the EEG signal as the output
random signal of a linear time invariant filter, where the input is white noise with a mean of zero and a
certain variance of σ2. The aim of the AR procedure is to obtain the filter coefficients, since it is
assumed that different thinking activities will produce different filter coefficients. The filter
coefficients will be used as the features of the signal.
AR assumes that the transfer function of the filter will only contain poles in the denominator. The
number of poles in the denominator corresponds to the order of the autoregressive model. The
assumption of an all-pole filter makes the filter coefficients computation easier because it is only
necessary to solve linear equations.
Mathematically, the AR model of order p describes the EEG signal y(t) as:
1 2 3 (5)
where, ai is the i-th filter coefficient, and n(t) is the noise. There are several methods that compute the
filter coefficients such as the Yule-Walker, Burg, covariance, and forward-backward algorithms [169].
The resulting coefficients can be used to estimate the power spectrum of the EEG signal y(ω),
such that:
1
(6)
1 ∑
where, ak are the estimated filter coefficients, and p is the AR model order, in other words, the number
of poles.
In the AR model, the determination of an appropriate order p for a given input signal is a trade-off
issue. If the order is too low to model the input signal, the result will not faithfully represent the signal
Sensors 2012, 12 1233
because the spectrum is too smooth. In contrast, if the order is too high, the spectrum may exhibit
spurious peaks.
AR spectral estimation is preferred to Fourier Transform, because of its superior resolution for short
time segments [170]. Nevertheless, AR performs poorly when the signal is not stationary [171]. Due to
the non-stationary nature of EEG signals, a multivariate adaptive AR (MVAAR) model has been
proposed to design more effective on-line BCI systems. Jiang et al. [172] applied MVAAR for the
classification of motor imagery, showing that MVAAR is a valuable adaptive method for feature
extraction. The computation algorithm was very similar to the original AR model. In a BCI with m
channels, the vector of m EEG values, at each point in time k, was represented as:
, , … , (7)
As in the AR case, the MVAAR model was expressed as:
(8)
where, was the vector of white noise values, were the adaptive coefficients, and p was the
model order. The Recursive Least Squares algorithm, a special variant of the Kalman Filter, were used
to update coefficients at every point k.
MF is a feature extraction method that attempts to detect a specific pattern on the basis of its
matches with predetermined known signals or templates. The intention of the user is revealed by
means of the correlation between the unknown EEG signals and the set of templates. Each template
represents an intention of the user. A higher correlation would imply better matching between the
template and the user’s intention. Each matched filter can simply be modeled as a sum of the
harmonically related sinusoidal components [151]:
2
cos (9)
where, n is the template sample number, fs is the sampling frequency, fF is the fundamental frequency
of the rhythm template, N-1 is the number of harmonics, and at and are the amplitude and phase of
the individual harmonics, respectively. The model parameters at and can be obtained from the FFT
spectrum [151].
MF has been proven especially effective for the detection of waveforms with consistent temporal
characteristics. Krusienski et al. [151] used MF for the identification of user intentions through µ-
rhythms and Brunner et al. [173] also used it for SSVEP feature extraction.
WT is a mathematical tool widely used for extracting information from many different kinds of
data, such as audio or image data, among others. WT is particularly suitable when signals are not
stationary, because it provides a flexible way of representing the time-frequency of a signal [174].
Sensors 2012, 12 1234
Wavelets are functions of varying frequency and limited duration that allow simultaneous study of
the signal in both the time and the frequency domain [175], in contrast to other modalities of signal
analysis such as Fourier transform (FT). FT provides only an analysis of the signal activity in the
frequency domain. FT gives information about the frequency content, but it is not accompanied by
information on when those frequencies occur. Short-term Fourier Transform (STFT) was proposed to
overcome this shortcoming of the Fourier analysis. The STFT divides the signal into successive time
windows and applies the FT in each epoch of the signal in time. In this approach, the design of window
length is a trade-off because smaller windows lead to higher temporal resolution but also to lower
frequency resolution at the same time. The WT overcomes this drawback by decomposing the signal in
both the time and the frequency domain at multiple resolutions, by using a modulated window that is
shifted along the signal at various scales.
Continuous wavelet transform (CWT) is defined as the convolution of the signal with the
wavelet function , [175]:
, , (10)
, is the wavelet coefficient that corresponds to the frequency associated with the scale s and the
time τ of the wavelet function , , and the symbol ‘*’ expresses the complex conjugation. The
wavelet function , is a dilated and shifted version of a mother wavelet :
1
, (11)
√
A mother wavelet can take multiples shapes, but it always satisfies the next condition:
0 (12)
The CWT defined in the Equation (10) is actually a kind of template matching, similar to a matched
filter in which the cross variance between the signal and a predefined waveform is calculated [151].
The advantage of the CWT over classic template matching methods arises from the special properties
of the wavelet template. The wavelets are suitable for transient signal analysis, in which the spectral
properties of the signal vary over time [176].
WT is a powerful tool for the decomposition of transient brain signals into their constituent parts,
based on a combination of criteria such as frequency and temporal position. Signals of identical
frequency ranges can be distinguished by means of the temporal position. Likewise, it is possible to
separate temporally overlapping processes thanks to the different frequency content.
The CWT introduces a lot of redundancy and complexity since it involves the analysis of a signal at
a very high number of frequencies using multiple dilations and shifting of the mother wavelet. Discrete
wavelet transform (DWT) was introduced to reduce this redundancy and complexity. The DWT
translates and dilates the mother wavelet in certain discrete values only [177]. Farina et al. [178]
showed a pattern recognition approach for the classification of single trial movement-related cortical
potentials, where the feature space is built from coefficients of a discrete wavelet transformation.
Although DWT is less redundant and less complex than CWT, CWT is still employed to extract
Sensors 2012, 12 1235
features from P300 and SCP, because it can clarify subtle information that DWT is unable to
extract [179].
The use of WT requires the selection of a mother wavelet. Many different mother wavelets can be
found in BCI applications and the selection of any one depends on what types of features need to be
extracted from the signal. The Mexican Hat wavelet is well localized in the time domain and is
employed for the localization of ERP components in time [179]. The Morlet wavelet is well localized
in the frequency domain and has been used for the analysis of gamma activity [180]. The bi-scale
wavelet has been employed successfully for designing an asynchronous BCI based on detection of
imaginary movement in the 1–4 Hz frequency range [181]. Also, the Daubechies wavelet, a very
well-known mother wavelet, has been used for the classification of SCPs [182].
CSP is a feature extraction method that projects multichannel EEG signals into a subspace, where
the differences between classes are highlighted and the similarities are minimized. It aims to make the
subsequent classification much more effective, by designing a spatial filter that transforms the input
data into output data with an optimal variance for the subsequent discrimination [183]. CSP has been
designed for the analysis of multichannel data belonging to 2-class problems. Nevertheless, some
extensions for multiclass BCIs have also been proposed [184].
CSP calculates the normalized spatial covariance C from the input data E, which represents the raw
data of a single trial, by means of:
(13)
where, E is an matrix, in which T is the number of channels, i.e., recording electrodes, and N the
number of samples per channel. The apostrophe denotes the transpose operator, while trace(x) is the
sum of the diagonal elements of x.
Assuming CSP is used to classify two classes, e.g., left and right motor imagery, CSP calculates the
spatial covariances and for each of the two classes by averaging the covariances over the
successive training trials of each class over time. The composite spatial covariance is computed as:
(14)
Since is real and symmetric, it can be factored as , where is the matrix of
eigenvectors, and is the diagonal matrix of eigenvalues.
By means of the whitening transform:
(15)
the variances are equalized in the space spanned by and all eigenvalues of are equal to one. If
and are transformed as:
(16)
(17)
then, and will share common eigenvectors. If , then , and ,
where is the identity matrix. As a result of the sum of two corresponding eigenvalues being always
Sensors 2012, 12 1236
one, the eigenvectors with the largest eigenvalues for correspond to the smallest eigenvalue for ,
and vice versa. This property is very useful for subsequent classification, because the variance of the
signal is maximized for one class while minimized for the other class.
Finally, the feature vector Z is obtained from the trial E as:
(18)
where, is the spatial filter matrix built by the CSP procedure.
CSP increases the accuracy of synchronous BCIs where it is allowed to send signals only during
certain predefined time periods. However, CSP does not offer the same improvement in asynchronous
BCIs. This is mainly due to the nonstationary properties of EEG signals [185]. Also, the performance
of CSP is affected by the spatial resolution, and it has been proven that some electrode locations offer
more discriminative information for some specific brain activities than others. For these reasons,
several methods improving the original CSP method have been proposed to increase the performance:
Wavelet Common Spatial Pattern (WCSP) [185], Common Spatio-Spectral Pattern (CSSP) [186], and
Common Sparse Spectral Spatial Pattern (CSSSP) [187].
GA is an optimization procedure to establish whether a certain set of features is the most efficient.
GA has been used in very diverse fields to solve optimization problems. In BCI research, GA has been
used as an automatic method to extract an optimal set of relevant features [188,189].
The baseline of the algorithm is a population of candidate solutions called individuals, creatures, or
phenotypes which are encoded by strings named chromosomes or the genotype of the genome. These
strings are coded either by binary information or no binary information. The standard steps of the GA
can be explained briefly as follows (Figure 3). GA begins with an initial population which is randomly
generated unless the algorithm has previous of the final solution. In the case of having initial
information, the initial population may be directed towards areas where optimal solutions are more
likely to reduce the number of iterations. The fitness of every individual population is evaluated.
According to their fitness, some representatives of the population may be discarded to vacate space for
newly generated individuals. Other individuals may be selected as parents in order to breed new
individuals. Also, some individuals may be stochastically selected to keep diversity in the population
preventing premature convergence. After the selection step, the individuals are crossed with each
other. In the crossover step, mating is performed among the selected parents to generate one or more
offspring. To keep a fixed population size, the number of offspring is usually the same as the number
of discarded individuals. The parents’ genes are split into pieces and then combined to form new
offspring. Following the crossover step, mutations are introduced to alter the population in order to
avoid converging towards a local suboptimum solution before exploring the entire search space. As a
result of the mutation, it is possible to discover areas that cannot be explored by crossover. Finally, the
fitness of the new population is evaluated. When an acceptable solution is reached or the maximum
number of generations has been produced, the algorithm is terminated. Otherwise, another iteration of
the algorithm is produced.
Sensors 2012, 12 1237
Start
Evaluation Crossover
Mutation
Optimum population No
or maximum number of
iterations?
Yes
End
Sequential selection is an optimization approach that aims at finding the optimal subset of features
by adding or removing features sequentially. There are two algorithms that perform sequential
selection: sequential forward selection and sequential backward selection.
Sequential forward selection (SFS) [190] is a bottom up algorithm. Firstly, the best individual
feature is found as the first feature in the subset. Next, for each subsequent step, the algorithm chooses
the feature from the remaining set, which in combination with the previously selected features, yields
the best subset of features. Finally, the algorithm finishes when the required number of features is
reached. The shortcoming of this algorithm is that the superfluous features are not removed once other
features are added. Sequential backward selection (SBS) [190], in contrast to SFS, is a top down
process. The process starts with the entire set of features and removes step by step features in such a
way that the error is as low as possible. This algorithm is also suboptimal, because it discards some
features that may be helpful after discarding other features. SFS has been used with success in the field
of BCIs [191,192].
Another refined method is introduced to partially overcome the aforementioned deficiencies. This
method, known as plus l take away r method (l > r), adds l features, and remove r features that is not
working well with other selected features. Sequential forward floating search (SFFS) or sequential
backward floating search (SBFS) are based on the plus l-take away r method [193]. SFFS starts with a
null feature set and, for each step, the r best features are included in the current feature set. In other
words, r steps of SFS are performed. Next, the algorithm verifies the possibility that some feature may
be excluded. Then, l worst features are eliminated from the set; in other words, l steps of SBS. SFFS
increases and decreases the number of features until the desired number of features is reached. SBFS
works analogously, but starting with the full feature set and performing the search until the desired
dimension is reached, using SBS and SFS steps.
In BCI research, SFFS has been used to reduce the dimensionality of the feature space to an
appropriate size for the available training data [194–196].
Sensors 2012, 12 1238
6. Artifacts in BCIs
Artifacts are undesirable signals that contaminate brain activity and are mostly of non-cerebral
origin. Since the shape of neurological phenomenon is affected, artifacts may reduce the performance
of BCI-based systems. Artifacts may be classified into two major categories: physiological artifacts
and non-physiological or technical artifacts.
Physiological artifacts are usually due to muscular, ocular and heart activity, known as
electromyography (EMG), electrooculography (EOG), and electrocardiography (ECG) artifacts
respectively [197]. EMG artifacts, which imply typically large disturbances in brain signals, come
from electrical activity caused by muscle contractions, which occur when patients are talking, chewing
or swallowing. EOG artifacts are produced by blinking and other eye movements. Blinking makes
generally high-amplitude patterns over brain signals in contrast to eye movements which produce
low-frequency patterns. These electrical patterns are due to the potential difference between the cornea
and the retina, as their respective charges are positive and negative. For that reason, the electric field
around the eye changes when this dipole moves. EOG artifacts mostly affect the frontal area, because
they are approximately attenuated according to the square of the distance [198]. Finally, ECG artifacts,
which reflect heart activity, introduce a rhythmic signal into brain activity [197].
Technical artifacts are mainly attributed to power-line noises or changes in electrode impedances,
which can usually be avoided by proper filtering or shielding [197]. Therefore, the BCI community
focuses principally on physiological artifacts, given that their reduction during brain activity
acquisition is a much more challenging issue than non-physiological artifact handling.
Several ways of handling physiological artifacts can be found in the literature. Artifacts may be
avoided, rejected or removed from recordings of brain signals. Artifact avoidance involves asking
patients to avoid blinking or moving their body during the experiments [199]. This approach to artifact
handling is very simple, because it does not require any computation as brain signals are not assumed
to have artifacts. However, this assumption is not always feasible given that some artifacts -
involuntary heart beats, eye and bodily twitches- are not easily avoidable during data recording,
especially in cases of strong neurological disorders [199]. Artifact rejection approaches suggest
discarding the epochs contaminated by the artifacts. Manual artifact rejection is an option to remove
artifacts in brain signals and an expert could identify and eliminate all artifact-contaminated epochs.
The main disadvantage in using manual rejection is that it requires intensive human labor, so this
approach is not suitable for on-line BCI systems. Nevertheless, this task can be performed
automatically by EMG and EOG artifact detection. If EMG and EOG signals are monitored, the brain
signal samples may be removed whenever ocular or muscular activity of the arms is detected [200].
Automatic rejection is an effective way of artifact handling, but it may fail when EOG amplitudes are
too small. Besides, rejection methodology means that the user loses device control when artifact
contaminated signals are discarded. Instead of rejecting samples, the artifact removal approach
attempts to identify and remove artifacts while keeping the neurological phenomenon intact. Common
methods for removing artifacts in EEG are linear filtering, linear combination and regression, BSS and
PCA [197]; some of which were discussed in Section 0.
Instead of avoided, rejected or removed artifacts from recordings of brain signals, some systems
acquire and process artifacts to offer a communication path that either disabled or healthy people can
Sensors 2012, 12 1239
use in many tasks and in different environments. This kind of system is not considered a BCI, because
communication is not independent of peripheral nerves and muscles. EMG computer interface [201],
human-computer interface (HCI) [202], EMG-based human-computer interface [203], EMG-Based
Human-Machine Interface [204], EMG-based human-robot interface [205], muscle-computer interface
(MuCI) [206], man-machine interface (MMI) [207], and biocontroller interface [208] are different
terms used to name communication interfaces in the scientific literature that can employ artifact
signals, among others. These systems usually have greater reliability than BCIs, but they cannot be
used by severely disabled people with strong constraints in voluntary movements.
7. Classification Algorithms
The aim of the classification step in a BCI system is recognition of a user’s intentions on the basis
of a feature vector that characterizes the brain activity provided by the feature step. Either regression or
classification algorithms can be used to achieve this goal, but using classification algorithms is
currently the most popular approach [209].
Regression algorithms employ the features extracted from EEG signals as independent variables to
predict user intentions. In contrast, classification algorithms use the features extracted as independent
variables to define boundaries between the different targets in feature space. McFarland et al. [210]
illustrated the differences between the two alternatives. For a two-target case, both the regression
approach and the classification approach require the parameters of a single function to be determined.
In a four-target case, assuming that the targets are distributed linearly, the regression approach still
requires only a single function. In contrast, the classification approach requires the determination of
three functions, one for each of the three boundaries between the four targets. Therefore, the
classification approach might be more useful for two-target applications and the regression approach
may be preferable for greater numbers of targets, when these targets can be ordered along one or more
dimensions. Moreover, the regression approach is better for continuous feedback e.g., applications
which involve continuous control of cursor movement. Figure 4 illustrates the differences between
classification and regression approaches.
Classification Regression
Sensors 2012, 12 1240
Classification algorithms can be developed via either offline, online or both kinds of sessions. The
offline session involves the examination of data sets, such as BCI competitions data sets [211], which
are collected from an adaptive or closed-loop system. The statistics of the data may be estimated from
observations across entire sessions and long-term computations may be performed. The results can be
reviewed by the analyst with the aim of fine-tuning the algorithms. Offline data analysis is valuable,
but it does not address real-time issues. In contrast, online sessions provide a means of BCI system
evaluation in a real-world environment. The data are processed in a causal manner and the algorithms
are tested in an environment in which the users change over the time as a result of e.g., changes in
motivation or fatigue. Although some researchers test new algorithms with only offline data, both
offline simulation and online experiments are necessary for effective algorithm design in closed-loop
systems. In other words, offline simulation and cross-validation can be valuable methods to develop
and test new algorithms, but only online analysis can yield solid evidence of BCI system
performance [137,212,213].
Classification algorithms have traditionally been calibrated by users through supervised learning
using a labeled data set. It is assumed that the classifier is able to detect the patterns of the brain signal
recorded in online sessions with feedback. However, this assumption results in a reduction in the
performance of BCI systems, because the brain signals are inherently non-stationary. In this regard,
Shenoy et al. [214] described two main sources of non-stationarity. On the one hand, the patterns
observed in the experimental samples during calibration sessions may be different from those recorded
during the online session. On the other hand, progressive mental training of the users or even
changes in concentration, attentiveness, or motivation may affect the brain signals. Therefore, adaptive
algorithms are essential for improving BCI accuracy. Adaptation to non-stationary signals is
particularly necessary in asynchronous and non-invasive BCIs [215,216].
Apart from the fact that supervised learning is not optimal for non-stationary signals classification,
large data sets and, thus, long initial calibration sessions are usually required to achieve acceptable
accuracy. Semi-supervised learning has been suggested to reduce training time and to update the
classifier in the online session on a continuous basis [217]. In semi-supervised learning, the classifier is
initially trained using a small labeled data set, after which the classifier is updated with on-line test data.
In a realistic BCI scenario, the signal associated with the subject’s intentions is not usually known
and the labels are not available. Either unsupervised learning or reinforcement learning can be applied
for BCI adaptation when the labeled data set is not available. Unsupervised methods attempt to find
hidden structures in unlabeled data, in order to classify them. Some unsupervised methods rely on
techniques for co-adaptive learning of user and machine [218,219] or covariate shift adaptation [220].
Reinforcement learning methods are based on the fact that distinguishing EEG potentials are elicited
when a subject is aware of an erroneous decision. These potentials are used as learning signals to
prevent that error from being repeated in the future [221].
The adaptation generally results in enhanced performance. Nevertheless, it is worth highlighting
that inherent risks exist in an adaptive BCI. A BCI that learns too fast may confuse the user, because
training will take place in a changing environment [222]. In addition, adaptive procedures can hide
some relevant signal features. Accordingly, there is a tradeoff between highly sensitive adaptation and
feature extraction.
Sensors 2012, 12 1241
Classifiers also have to face two main problems related to the pattern recognition task: the curse of
dimensionality and the bias-variance tradeoff. The curse of dimensionality means that the number of
training data needed to offer good results increases exponentially with the dimensionality of the feature
vector [223]. Unfortunately, the available training sets are usually small in BCI research, because
training process takes a long time and is a tiring process for users. The bias-variance tradeoff
represents the natural trend of the classifiers towards a high bias with low variance and vice versa.
Stable classifiers are characterized by high bias with low variance, while unstable classifiers show high
variance with low bias. To achieve the lowest classification error, bias and variance should be low
simultaneously. A set of stabilization techniques such as the combination of classifiers or regularization
can be used to reduce the variance.
The design of the classification step involves the choice of one or several classification algorithms
from many alternatives. Several classification algorithms have been proposed such as k-nearest
neighbor classifiers, linear classifiers, support vector machines, and neural networks, among others.
The general trend prefers simple algorithms to complex alternatives. Simple algorithms have an
inherent advantage because their adaptation to the features of the brain signal is inherently simpler and
more effective than for more complex algorithms. Nevertheless, simple algorithms, whenever
outperformed in online and offline evaluations, should be replaced by more complex alternatives [213].
Finally, certain inherent dangers of classification algorithm usage should be pointed out. Although
classification algorithms have clearly helped to characterize task relevant brain states, several pitfalls
may occur when these algorithms are used by non-experts. Bias and variance of the estimated error of
the algorithms, and their overfitting are the main source of difficulties [224]. If a classifier is
overfitted, then it will only be able to classify the training data or similar data. Overfitting can be
avoided by restricting the complexity of the classification procedure [224]. Classification error is
estimated by means of cross validation. Once a classification algorithm is trained, the algorithm is
validated on a validation data set, which should be independent of the training data set. This procedure
is usually repeated several times, using different partitions of the sample data. The resulting validation
errors are averaged across multiple rounds. This approach presents some inherent dangers that must be
prevented, because some elements of the partition may not be independent of each other or may not be
identically distributed, among other reasons [224].Next, this section presents the properties of a set of
classifiers, in order to make it easier to choose an appropriate classifier for a given type of BCI. All
classifier methods are listed in Table 8, along with their main properties.
K-nearest neighbor classifiers (k-NNC) are based on the principle that the features corresponding to
the different classes will usually form separate clusters in the feature space, while the close neighbors
belong to the same class. This classifier takes k metric distances into account between the test samples
features and those of the nearest classes, in order to classify a test feature vector. The metric distances
are a measure of the similarities between the features of the test vector and the features of each class.
The advantage of taking k neighbors into account in the classification is that error probability in the
decision is decreased. Some training samples may be affected by noise and artifacts, which may
seriously influence the classification results. If a decision involving several neighbors is made, then it
is less likely that an error will occur, because the probability of several simultaneous erroneous datum
is much lower [225].
Rather than only the closest sample, if several k closest classes are considered, then a voting scheme
is required to decide between competing choices. Since there are no reasons to assume that the
distributions of those neighbors are homogenous, it is clear to see that the k-NNC has to assign
different ranks to the nearest neighbors, according to their distances from the test example. Therefore,
k-NNC needs to define a weighting function, which varies with the distance in such a way that the
output value decreases as the distance between the test feature vector and the neighbor increases. The
function defined by Equation (19) [226] meets this requirement:
(19)
1
where, denotes the distance of the i-th nearest neighbor from a test example. That is,
corresponds to the nearest neighbor and to the furthest. The decision rule of k-NNC assigns the
unknown examples to the class with the greatest sum of weights among its k nearest neighbors.
k-NNC is not very common in BCI research, because this classifier is very sensitive to the
dimensionality of the feature vector [227]. Nevertheless, k-NNC has been proven to be efficient with
Sensors 2012, 12 1243
low dimension feature vectors. Also, k-NNC has been tested in a multiclass environment [228] and
applied to cursor movements on a vertical axis, when classifying SCPs [229].
LDA is a very simple classifier that provides acceptable accuracy without high computation
requirements. LDA is very common in the BCI community and is a good choice for designing online
BCI systems with a rapid response, but limited computational resources. LDA provides relatively
acceptable accuracy and has been used successfully in numerous BCI systems, such as P300
speller [179], multiclass [230], or synchronous [231] BCIs. Nevertheless, it can lead to completely
erroneous classifications in the presence of outliers or strong noise [232]. LDA is usually applied to
classify patterns into two classes, although it is possible to extend the method to multiples classes [230].
For a two-class problem, LDA assumes that the two classes are linearly separable. According to this
assumption, LDA defines a linear discrimination function which represents a hyperplane in the feature
space in order to distinguish the classes. The class to which the feature vector belongs will depend on
the side of the plane where the vector is found (Figure 5). In the case of an N-class problem (N > 2),
several hyperplanes are used. The decision plane can be represented mathematically as:
(20)
where, w is known as the weight vector, x is the input feature vector and w0 is a threshold. The input
feature vector is assigned to one class or the other on the basis of the sign of .
Figure 5. Linear classifier and margins. The decision boundary is the thick line. (adapted from [232]).
There are many methods to compute w. For example, w may be calculated as [233]:
μ μ (21)
where, μi is the estimated mean of class i and ( is the estimated common covariance
matrix; the average of the two class empirical covariance matrices. The estimators of the covariance
matrix and of the mean are calculated as:
1
μ μ (22)
1
Sensors 2012, 12 1244
1
μ (23)
Figure 6. Eigenvalue spectrum of a given covariance matrix (bold line) and eigenvalue
spectra of covariance matrices estimated from a finite number of samples (N = 50, 100, 200,
500). Note that accuracy increases as the number of trials increase (adapted from [233]).
250
real
N=500
200 N=200
N=100
N=50
Eigenvector
150
100
50
0
0 50 100 150 200
Index of eigenvalue
The estimation of the covariance defined in Equation (22) is unbiased and has good properties
under usual conditions. Nevertheless, it may become imprecise in some cases where the dimensionality
of the features is too high compared to the number of available trials. The estimated covariance matrix
is different from the true covariance matrix, because the large eigenvalues of the original covariance
matrix are over estimated and the small eigenvalues are under estimated (Figure 6). It leads to a
systematic error which degrades LDA performance [233].
For this reason, a new procedure has been proposed to estimate the covariance, improving the
standard estimator defined in the Equation (22). The new standard estimator of the covariance matrix
is given by:
1 (24)
The γ value is referred to as a shrinkage parameter and is tunable between 0 and 1. ν is defined as
/ with d being the dimensionality of the features space. The selection of a shrinkage
parameter implies a trade-off and is estimated on the basis of the input data [234].
Some improved algorithms have been introduced based on LDA such as Fisher LDA (FLDA) and
Bayesian LDA (BLDA) [235]. In the first example, performance was improved by projecting the data
to a lower dimensional space, in order to achieve larger intervals between the projected classes and,
simultaneously, to reduce the variability of the data in each class. However, FLDA does not work well
Sensors 2012, 12 1245
when the number of features becomes too large in relation to the number of training examples. This is
known as the small sample size problem [235].
The second modification can be seen as an extension of FLDA. BLDA solves the small sample size
problem by introducing a statistical method known as regularization. The regularization is estimated
through Bayesian analysis of training data and is used to prevent overfitting of high dimensional and
possibly noisy datasets. Overfitting means the classifier has lost generality and is therefore undesirable
in a classifier. If a classifier is overfitted, then it is only able to classify the training data or similar data.
In comparison to FLDA, the BLDA algorithm provides higher classification accuracy and bitrates,
especially in those cases where the number of features is large [235]. Additionally, BLDA requires
only slightly more computation time, which is a crucial requirement in real BCI systems.
SVM is a classifier that, in a similar way to LDA classifiers, constructs a hyperplane or set of
hyperplanes, in order to separate the feature vectors into several classes. However, in contrast to LDA,
SVM selects the hyperplanes that maximize the margins, that is, the distance between the nearest
training samples and the hyperplanes [236]. The basis of SVM is to map data into a high dimensional
space and find a separating hyperplane with the maximal margin [237] according to Cover’s theorem
on the separability of patterns [238]. Cover’s theorem states that a complex classification problem cast
in a high-dimensional nonlinear space is more likely to be linearly separable than in a low-dimensional
nonlinear space. Also, as for linear analysis classifier, an SVM uses regularization, in order to prevent
the classifier from accommodating possibly noisy datasets.
SVM has been used to classify feature vectors for binary [239,240] and multiclass problems [228,230].
It has also been successfully used in a large number of synchronous BCIs [131,230,240]. Such a
classifier is regarded as a linear classifier, since it uses one or several hyperplanes. Nevertheless, it is
also possible to create a SVM with non-linear decision boundary by means of a kernel function K(x, y).
Non-linear SVM leads to a more flexible decision boundary in the data space, which may increase
classification accuracy. The kernel that is usually used in the BCI field is the Gaussian or Radial Basis
Function (RBF):
, exp (25)
2
The Gaussian SVM has been applied in BCIs to classify P300 evoked potentials [241–243].
SVM has been widely used in BCI, because it is a simple classifier that performs well and is robust
with regard to the curse of dimensionality, which means a large training set is not required for good
results, even with very high dimensional feature vectors [228]. These advantages come at the expense
of execution speed. Nevertheless, SVM is speedy enough for real-time BCIs [243,244].
Bayesian statistical classifier is a classifier which aims to assign, with the highest probability, an
observed feature vector x from its class y. The Bayes’ rule is used to obtain the a posteriori probability
| that a feature vector has of belonging to a given class. Assuming, for example, two classes L
Sensors 2012, 12 1246
and R corresponding to imaginary left and right movements of the hand, the a posteriori probabilities
of each class are computed using the Bayes’ rule as:
| | |
| (26)
| | | |
Typically, it is assumed that the a priori probabilities are equal, P(y) = P(L) = P(R) = 0.5, since it
is supposed the user has no predilection for any movement. In order to calculate the probabilities
| , it is usually supposed that a Gaussian statistical distribution applies to the features for each
class, although it may also be assumed that the distribution is a weighted mixture of Gaussian
distributions [245]:
| | (27)
where, wi is the weight of each Gaussian prototype and M is the number of prototypes. Two ways are
feasible to estimate the Gaussian prototypes mixture [245]. The first is to divide the feature space in
several equally sized regions and calculate the mean and variance of the Gaussian prototypes in each
area from training data. The set of Gaussian prototypes is equally weighted and the weights wi are
equal to . The second uses a Gaussian mixture models (GMM). The different weights wi and the
mean, variance, and covariance matrices that define each Gaussian prototype, are calculated by the
expectation maximization (EM) algorithm. EM algorithm is an iterative procedure which guarantees
the maximum likelihood or maximum a posteriori (MAP) estimates of the parameters in the statistical
model. Lui et al. [246] made GMM adaptive to significant changes in the statistical distribution of the
data during long-term use. In these improvements, the initial mean, variance and covariance of each
class is updated over time using a specific number of recent trials.
Bayesian statistical classifiers are not very popular in the BCI community. Nevertheless, they have
been used for classifying motor imagery [247] or visual P300 evoked potentials [248].
ANNs are non-linear classifiers that have been used in many applications, in a wide variety of
disciplines such as computer science, physics, and neuroscience. The idea of ANNs is inspired in how
the brain processes the information. The purpose is to mimic brain activity that immediately solves
certain problems, which a conventional computer program processes poorly. For example, ANNs are
widely used in pattern recognition, because they are capable of learning from training data. The ability
to learn from examples is one of most important properties of ANNs. Once trained, the ANNs are
capable of recognizing a set of training data-related patterns. ANNs are therefore associated with BCI
applications, since pattern recognition is performed to ascertain user intentions.
An ANN comprises a set of nodes and connections that are modified during the training process.
The ANN is fed on a set of training examples and the output is observed. If the output is incorrect, then
the internal weights are modified by the training algorithm to minimize the difference between desired
and actual output. This training continues until the network reaches a steady state, where no further
Sensors 2012, 12 1247
significant improvement is achieved. In this state, not only should the ANN produce correct outputs for
all examples of the training set, but also for inputs that were not encountered during training.
From a mathematical point of view, ANNs define a mapping from an input space to an output
space, that can be described as a vector-valued function , where both x and y may be of any
dimensionality. The mapping function f is a combination of mappings, which are individually
performed by single nodes or neurons. Each neuron processes the information non-linearly and the
resulting mapping is therefore non-linear. This property is important, especially in those cases where
the physical mechanism that generates the input signal is non-lineal.
One of the most well-known ANN structures is the multilayer perceptron (MLP) introduced by
Rumelhart and McClelland in 1986. MLPs are very flexible classifiers that can classify any number of
classes and adapt to numerous kinds of problems. In the field of BCIs, MLP have been applied to
classify two [249], three [200], and five [250] different tasks, and to design synchronous [251] and
asynchronous [215] BCIs. Moreover, MLP has been used for preprocessing EEG signals before the
feature extraction step rather than the classification step, in order to improve the separability of EEG
features [252].
Besides MLP, different types of ANN architecture have been used in the design of BCI systems
such as Probabilistic Neural Networks (PNN) [253, 254], Fuzzy ARTMAP Neural Networks [255],
Finite Impulse Response Neural Networks (FIRNN) [251] or Probability estimating Guarded Neural
Classifiers (PeGNC) [256].
8. BCI Applications
BCIs offer their users new communication and control channels without any intervention of
peripheral nerves and muscles. Hence, many researchers focus on building BCI applications, in the
hope that this technology could be helpful for those with severe motor disabilities. Various BCI
applications have very recently been developed thanks to significant advances in the field of
EEG-based BCI. EEG signals are used by most BCI applications, because they offer an acceptable
signal quality that combines low cost and easy-to-use equipment. Thanks to BCI applications, it is
hoped that the quality of life of severely disabled people can be improved. Likewise, the attention
given by caregivers will be less intensive, reducing its costs and making the life of relatives less
onerous. Moreover, BCI applications potentially represent a powerful tool for revealing hidden
information in the user’s brain that cannot be expressed.
The main target populations for BCI applications fall into three classes. The first group includes
Complete Locked-In State (CLIS) patients who have lost all motor control, because they may be at a
terminal stage of ALS or suffer severe cerebral palsy. The second group comprises Locked-In State
(LIS) patients who are almost completely paralyzed, but with residual voluntary movement, such as
eye movement, eye blinks, or twitches with the lip. The third group of potential BCI users includes
abled bodied people and those with substantial neuromuscular control, particularly speech and/or hand
control. BCI have little to offer to the third group, because they can send the same information much
more quickly and easily via other interfaces, rather than a BCI. Despite this, BCIs are increasingly
used by healthy people in neuromarketing and video games as a tool to reveal affective information of
Sensors 2012, 12 1248
the users, which cannot be so easily reported through conventional interfaces. Likewise, BCI can be
used for some people that suffer from neurological disorders such as schizophrenia or depression.
The level of impairment of the potential target population is related to the performance of a BCI
system. Kübler et al. [257] reported a strong correlation between physical impairment and BCI
performance. CLIS patients were unable to control a BCI. Voluntary brain regulation for communication
was only possible in LIS patients. However, considering only LIS patients, this relationship between
physical impairment and BCI performance disappeared. Figure 7 shows the relationship between BCI
application areas and BCI information transfer rates and user capabilities.
Figure 7. Relationship between BCI application areas, BCI information transfer rates and
user capabilities. Horizontal axis: information transfer rate that would make the application
controllable. Vertical axis: the degree of capability.
people Entertainment
Neuroprosthesis
LIS patients Locomotion
Environmental
control
Communication
CLIS patients
It is currently unclear whether BCI technology will ever outperform other established technologies
that include eye or muscle-based devices. Currently the latter devices tend to be easier to use and offer
better benefit/cost ratios [258,259]. For example, the detection of eye movement is quicker, easier, and
more accurate than the detection of ERP modulations. A spelling rate of 10 words per minute can be
obtained with unimpaired eye movement, by means of an eyetracker [260]. In that regard, hybrid BCI
systems have been proposed to improve performance. They are the combination of two different kinds
of BCIs or the combination a BCI with other existing assistive technology [261]. Unless the
performance of BCI systems improves considerably, BCI as assistive technology may only be
especially attractive for severely disabled people, when other technologies are unsuitable.
At present, LIS patients and those likely to develop CLIS constitute the principal candidates for
BCI. Despite the low information transfer rates provided by BCI, the high grade of disability among
LIS patients force them to use a BCI rather than more reliable conventional interfaces, such as muscle
or eye-gaze based system. Eye-gaze control constraints in some LIS patients are an important issue,
because they are obliged to use BCIs that does not depend on eye-gaze control [262,263]. Also,
eye-gaze control constraints make some BCI applications more difficult, such as steering a wheelchair.
Nowadays, there are a vast number of very different BCI applications, such as word processors,
adapted web browsers, brain control of a wheelchair or neuroprostheses, and games, among others.
However, most applications have solely been designed for training or demonstration purposes. Despite
the most recent significant advances in BCI technology, there are still many challenges to employing
BCI control for real-world tasks [264]: (i) the information transfer rate provided by BCIs is too low for
Sensors 2012, 12 1249
natural interactive conversation, even for experienced subjects and well-tuned BCI systems; (ii) the
high error rate further complicates the interaction; (iii) BCI systems cannot be used autonomously by
disabled people, because BCI systems require assistants to apply electrodes or signal-receiving devices
before the disabled person can communicate; (iv) a BCI user may be able to turn the BCI system off by
means of brain activity as input, but usually cannot turn it back on again, which is termed the “Midas
touch” problem; and (v) handling BCI applications demands a high cognitive load that can usually be
achieved by users in quiet laboratory environment, but not in the real world. Nevertheless, despite all
these challenging difficulties, the first steps on the path to long-term independent home use of BCIs
have already been taken [12].
Before describing the practical usage of BCI applications, it is worth considering the distinction
between BCIs and their applications [8]. As a tool that executes a specific function, particular BCI
specifications correspond to the way it performs that function. These specifications can therefore be
applied to wide variety of applications, even though the function remains unchanged. The important
thing in BCI evaluation is its performance when executing its specific function. In contrast,
applications are described in terms of the tools they employ and the purposes they serve. Therefore,
BCI evaluation focuses on how well it performs its purpose. In other words, the term BCI refers to the
system that records, analyses, and translates the input into commands and the term application denotes
the environment in which the BCI estimated output commands are applied. Consequently, the
evaluation procedures for BCI systems and their applications differ in each case. The following
sub-sections briefly describe BCI applications, classified into five main areas: communication, motor
restoration, environmental control, locomotion and entertainment.
8.1. Communication
BCI applications for communication deal with severe communication disabilities resulting from
neurological diseases. This kind of application probably represents the most pressing research in the
field of BCI, because communication activity is essential for humans. Applications for communication
purposes outline an operation that typically displays a virtual keyboard on screen, where the user
selects a letter from the alphabet by means of a BCI. The distinguishing element in each approach is
usually the BCI and the type of control signal.
Voluntary control of SCPs may be used for letter selection. With extensive training, completely
paralyzed patients are able to produce positive and negative changes in their SCP to drive the vertical
movement of a cursor [117]. Based on this kind of control signal, Birbaumer et al. [265] developed a
spelling device with an on-screen display, which used a cursor to select letters of the alphabet. Trials
involving two patients at advanced stages of ALS showed that they achieved a rate of about
2 characters per minute when writing text messages. Other types of control signals, such as detection
of eye blinks [266], which normally represent an artifact in EEG signals, or classification of three
mental tasks [215], are also used to select the blocks or characters in a virtual keyboard. Both
approaches are nearly the same apart from the control signal. In both cases, the virtual keyboard
consisted of a total of 27 symbols, 26 English letters plus the space to separate words, organized in a
three row by nine column matrix. Likewise, both applications were based on the same protocol of
writing a single letter, which required three steps. Firstly, the whole keyboard was divided into three
Sensors 2012, 12 1250
blocks, each with nine letters each. Then, the user could select a set of nine letters by producing a
single, two or three eye blinks [266] or imagining one of three available tasks [215] depending on the
case. After the first selection, the set of nine letters was distributed into three subsets, each with three
letters, and once again the user again selected one of them. Finally, at the third level, the user chose a
single letter amongst the three remaining symbols. The correct spelling rate of each speller was one
character per minute using blinks [266] and 2.73 characters per minute for three mental tasks [215].
Obermaier et al. [267] also designed a letter spelling based on standard Graz-BCI which also
included a virtual keyboard. The letter selection protocol is very similar to the approaches discussed
above, except that the entire alphabet consisted of 32 letters and was divided into two halves at each
step. In this case, the user chooses either subset of letters by EEG modulation through mental hand and
leg motor imagery. The spelling rate achieved by three healthy users varied between 0.5 and 0.85 letters
per minute. This is a lower rate than in previous cases, nevertheless, it appears easy to increase the
number of letters spelled per minute just by expanding the number of classes to more than two.
P300 event-related brain potentials are also very popular in BCI letter spelling applications.
P300-based BCIs have been proven sufficiently suitable for ALS patients in the early and middle
stages of the disease [268]. Besides, this kind of BCI is very handy because the P300 response occurs
spontaneously and consequently does not require substantial training. Furthermore, recent progress
with P300-based spellers have allowed the development of commercial applications available to the
general public [269]. One of the best-known P300 spellers was designed by Farwell and Donchin in
1988 [123]. In this speller, the 26 letters of the alphabet, together with several other symbols and
commands, are displayed on-screen in a 6 × 6 matrix (Figure 8) with randomly flashing rows and
columns. Then, the user focuses attention on the screen and concentrates successively on the characters
to be written, while the EEG response is monitored. Two P300 are elicited for each looked-for element
on the matrix, when the desired row or column flashed, thereby allowing the system to identify the
desired symbol. The results of the Farwell-Donchin speller trials involving 4 healthy people yielded an
acceptable spelling rate of about 2 characters per minute.
Figure 8. Original P300 speller. Matrix of symbols displayed on a screen computer which
serves as the keyboard or prosthetic device (adapted from [123]).
The Farwell-Donchin speller provides a relatively high rate and accuracy, but its precision can be
improved by reducing perceptual errors in the Farwell-Donchin paradigm [270]. Perceptual error
happens when a P300 response is elicited due to flashing rows or columns adjacent to the target
symbol, an issue which is its major source of error. Hence, a new letter distribution was presented to
Sensors 2012, 12 1251
overcome this problem (Figure 9) [270]. The idea is to have several regions flashing instead of using
rows and columns. The characters are placed into a two-level distribution. At the first level, the
characters are distributed into seven groups, each with seven characters, which are also flashing
randomly. The group containing the target character is found by P300 detection. At the second level,
the characters in the detected group are repositioned and the level one procedure is repeated, and so on
until the target character is finally selected.
Figure 9. The proposed region-based paradigm for the improved P300 speller: (a) The first
level of intensification where each group contains up to seven characters; and (b) One
region is expanded at the second level (adapted from [270]).
A B H I
C DE J K L
F G M N
OP VW 3 4
Q RS X Y Z 5 6 7
T U 1 2 8 9
0 ~ ^ *
! @# ( ) _
$ % = +
(a) (b)
Townsend et al. [137] presented a newly enhanced BCI based on a checkerboard paradigm instead
of the standard row/column paradigm introduced by Farwell and Donchin. In this new approach, the
standard matrix containing the targets was superimposed on a checkerboard. Trials with advanced ALS
patients and healthy people showed a significantly higher mean accuracy for the checkerboard
paradigm than for the row/column paradigm. Ahi et al. [271] also recently improved the
Farwell-Donchin P300 speller by introducing a dictionary to decrease the number of misclassifications
in the spelling. The dictionary was used for checking the candidate word proposed by the classifier of
P300 responses. In case of misspelling, the dictionary gave a certain number of suggestions from
which the system could select. Additionally, in order to reduce the probability of misspelling due to
perceptual errors, the usual letter position in the matrix was changed according to the analysis of word
similarities in the constructed dictionary.
All previous P300 spellers are based on the recording of visual event-related brain potentials.
However, there is no sense in using visual stimuli in cases of severely paralyzed patients with impaired
vision or poor control over eye movements. In these cases, auditory stimulation is used in order to
make P300 spellers suitable for this group of patients [129,272–274].
Other important applications of communication-related BCIs are Internet browsers adapted to users
with severe disabilities because, over the last decade, the Internet has become a very important part of
daily life. In this area, “Descartes” is one of the first EEG-controlled web browsers which can be
operated by SCPs [275]. Its browser interface is based on arranging the links alphabetically in a
dichotomous decision tree, where the user selects or rejects each item, producing positive or negative
SCP shifts. “Descartes” presents the shortcoming that only a limited number of web pages can be
Sensors 2012, 12 1252
browsed, because the user receives a number of predefined links arranged in a tree at the start of the
web surfing. Besides, graphical links cannot be chosen since the textual label is used to identify the
link. A more advanced prototype, called “Nessi”, overcomes these shortcomings thanks to a better user
interface [276]. Colored frames are placed around links or selectable items on the web page instead of
arranging the links in a tree. More recently, evoked potentials are also used to enhance browser
functionality. Mugler et al. [128] built an Internet browser with P300 control where the options are all
presented as icons in an 8x8 matrix. Jinghai et al. [103] developed a browser based on VEPs. One of
the advantages of ERPs is that they occur quickly and can lead to relatively high web surfing speeds.
Spinal cord injury (SCI) or other neurological diseases with associated loss of sensory and motor
functions dramatically decrease the patient’s quality of life and create life-long dependency on home
care services. Motor restoration may alleviate their psychological and social suffering. Restoring
movement, such as grasping, is feasible in quadriplegic patients through neuroprostheses guided by
functional electrical stimulation (FES). FES compensates for the loss of voluntary functions by
eliciting artificial muscle contractions. Electrical currents generate artificial action potential by
depolarizing intact peripheral motor nerves that innervate the targeted muscle and cause a muscle
contraction (see [277] for a review). EEG-based BCI can be used to generate a control signal for the
operation of FES, because EEG signals are unaffected by electrical activation of upper extremity
muscles [278]. Thanks to their merging of BCI and FES, Pfurtscheller et al. [279] developed an
application where a tetraplegic patient, suffering from a traumatic spinal cord injury, was able to
control paralyzed hands to grasp a cylinder. In that application, the patient generated beta oscillations
in the EEG by foot movement imagery. Then, the BCI analyzed and classified the beta burst and the
output signal was used to control the FES device that activated the extremity. Also, FES has been used
for rehabilitation training after a stroke. Hu et al. [280] developed a combined FES-robot system which
was continuously driven by the user’s residual electromyography on the affected side for wrist joint
training after a stroke, in order to involve the user’s own neuromuscular effort during the training.
FES has been proven to be an effective way to restore movement. Nevertheless, FES requires the
use of residual movements, which are not possible in severely injured patients. For this reason, some
groups have started to explore approaches that couple neuroprostheses and BCI without FES
intervention. Pfurtscheller et al. [281] demonstrated that a tetraplegic patient, whose residual
upper-limb muscle activity was restricted to the left biceps, due to an upper spinal cord injury, could
effectively control a hand orthosis using changes in Rolandic oscillations, which were produced by
motor imagery. A lengthy training period was required to use this application. However, the patient
was finally able to open and close the hand orthosis almost without any errors. Some years later, the
same group validated the coupling of EEG-based BCIs and an implanted neuroprosthesis giving
further evidence that BCI is a feasible option for the control of a neuroprostheses [282]. In this study,
BCI classified distinctive EEG-patterns that involved power decreases in certain specific frequency
bands. These patterns were generated by the user from mental imagery of his paralyzed left hand in motion.
More recently, ERPs are also used to provide motor restoration. Muller et al. [13] presented a novel
neuroprosthetic device for the restoration of the grasp function for people spinal cord injuries. This
Sensors 2012, 12 1253
neuroprosthetic device consisted of a dual-axis electrical hand prosthesis controlled by BCI based on
four-class SSVEPs. Hence, it is possible to select only four movements according to the four LEDs
flickering in different frequencies. The user’s gaze shifted between the different LEDs in order to
select a movement. One light on the finger index flickering at 6 Hz and another light on the pinky
finger flickering at 7 Hz served to turn the hand in supination or pronation. The two remaining lights
on the wrist flickering at 8 Hz and 13 Hz represented the orders to open and close each hand.
Within the field of BCI application in motor restoration, BCI systems have been also applied for
movement reconstruction in patients with severe post-stroke motor disability. BCI training is
hypothesized to provide feedback to sensorimotor cortex and, by doing so, movement is restored as
cerebral pathways reorganize to link up motor commands with motor movements. Buch et al. [283]
developed a BCI system that used MEG activity evoked by patient intent to move a completely
paralyzed hand, in order to control grasping motions of a mechanical orthosis attached to the affected
hand. Thanks to the hand prosthesis attached to the paralyzed hand and using visual feedback, the
patient could learn to open a hand by increasing SMR over the injured hemisphere and to close the
hand by decreasing it. MEG provides a much larger and more localized SMR response, which means
that even a digit finger may be controlled [284].
MEG-based BCI is too expensive for widespread applications. For that reason, Broetz et al. [285]
proposed a combination of MEG and EEG-based BCIs. Initially, the MEG-based BCI was used to
boost rehabilitation training success. Later, the user continued rehabilitation training with an EEG-based
BCI; a more affordable technology than MEG. Finally, the patient practiced physiotherapy training.
The results of this study suggest that the combination of BCI training with goal-directed active
physical therapy improves the motor abilities of chronic stroke patients. In similar experiments,
Caria et al. [286] reaffirmed the success of a combination of BCI training and physiotherapy. This
study encourages further research on the role of BCIs in brain plasticity and post-stroke recovery.
One of the main goals of BCI-based applications is to achieve maximum independence for the
patient, despite any motor disability. People who suffer severe motor disabilities are often homebound
and for this reason, environmental control applications focus on the control of domestic devices such
as TV, lights or ambient temperatures. Apart from improving the quality of life of severely disabled
people, assistive devices mean that the tasks of the caregiver are less intensive, costs are reduced, and
the life of relatives is less onerous.
Cincotti et al. [14] presented a pilot study dealing with the integration of BCI technology into the
domestic environment. In this study, fourteen patients with severe motor disabilities, due to
progressive neurodegenerative disorders, tested a device that provided environmental control through
an interface designed to support different levels of motor capacities for each user. Typical peripherals
such as keyboard, mouse or joystick were offered to allow the device control through upper limb
residual motor abilities. Head trackers and microphones for voice recognition were also available in
cases of people with impaired limbs but intact neck muscles and comprehensive speech. Lastly, in
cases of totally disabled people, the system could be controlled by voluntary modulations of
sensorimotor rhythms recorded by the EEG-based BCI. Thereby, the application offered the patient
Sensors 2012, 12 1254
different access modalities that matched their gradual loss of motor abilities due to progressive
neurodegenerative diseases. As output devices, the system allowed the use of a basic group of
domestic appliances such as lights, TV and stereo sets, a motorized bed, an acoustic alarm, a front door
opener, and a telephone, as well as wireless cameras to monitor the surrounding environment.
Invasive techniques have also been proposed in environmental control applications. Hochberg et al. [287]
implanted BrainGate sensors in the primary motor cortex of a tetraplegic patient to control a cursor.
The initial trials yielded promising results, where the patient could handle e-mail applications or
operate devices such as a television by imagining limb motions, even while conversing.
8.4. Locomotion
BCI applications that allow disabled people to control a means of transportation represent an
important field in their use. Thanks to these applications, people suffering from paraplegia or with
other physical impairments can autonomously drive a wheelchair, making them more autonomous and
improving their life quality. Portability is a necessity for these kinds of applications. Hence the use of a
BCI based on EEG recording is enforced. EEG signals are typically very noisy and are highly variable,
which means a relatively long time between commands that will be of high uncertainty. Therefore, the
main challenge is to achieve sufficient accuracy in driving as well as reaching real-time control, in
spite of the ultra-low information transfer rates provided by BCI. For this reason, some studies on
BCIs proposed invasive techniques to record EEG signals, because they achieved more spatial
resolution and reduced noise. Serruya et al. [288] experimented with monkeys implanting an electrode
array in the motor cortex. These initial experiments showed that the monkeys were able to move a
computer cursor to any position, opening up new human applications.
However, the risks related to invasive BCIs lead research towards building non-invasive
applications for human use. Some pilot experiments concerned with locomotion illustrate the
feasibility of using EEG signals for continuous control of a mobile robot in an indoor environment
with several rooms, corridors, and doorways [215,289]. The results of these experiments opened the
possibility for physically disabled people to use a portable EEG-based BCI for controlling wheelchairs.
To the best of our knowledge, in 2005, Tanaka et al. [127] presented the first application of wheelchair
control using only EEG. In this study, the surrounding floor was divided into squares between which
the user decided to move by imagining left or right-limb movements. Driven by user decisions, the
wheelchair therefore moved from one square to another. Tests with six healthy subjects were quite
encouraging and demonstrated the viability of wheelchairs control solely through the use of EEG signals.
In wheelchair control by BCI-based systems, the usual problems are the infrequent control signal
and the low information transfer rate and accuracy provided by a BCI. In that respect, some
improvements have been presented over the past few years. Synchronous P300-based BCIs have been
introduced in order to assure better accuracy. Likewise, to overcome the usual low bit rate in BCIs, the
systems have been endowed with certain autonomy, decreasing the number of interactions required.
Rebsamen et al. [290] designed a simplified wheelchair control by constraining the movements to
guidepaths defined by the patient or a helper. These guidepaths were attached to a specific point in the
environment and stored by the system. The user selected the destination through a P300-based BCI and
the wheelchair automatically followed the path. The user only had to decide when the wheelchair
Sensors 2012, 12 1255
would stop. For path guidance, the system steering the wheelchair had to be kept informed of its
localization uninterruptedly. To that effect, the wheelchair relied on an odometer and a bar-code
scanner to read bar-code patterns previously placed on the floor along the paths. Some years later, the
system was improved to ensure safer control. Two faster BCIs based on P300 and the µ/β rhythm were
employed, allowing the user to stop the wheelchair more quickly [291]. Both applications were tested
with healthy people.
The main disadvantage found in the preceding approaches is that the control assistance has little
flexibility and is not capable of dealing with unknown and populated scenarios. Iturrate et al. [292]
overcame this shortcoming by making the system create a dynamic reconstruction of the surrounding
scenario. Other studies suggested that help should only be available in those cases where the user
experienced more difficulties driving the wheelchair e.g., in a narrow corridor [216,293,294]. Three
levels of assistance may be possible in the shared control: collision avoidance, obstacle avoidance and
orientation recovery, which are only activated as required by the user [293]. Before executing the
user’s steering commands, the share control evaluates the situation from the data provided by a set of
laser scanners. Scanners inspect the environment and detect potential obstacles or walls.
8.5. Entertainment
Entertainment-orientated BCI applications have typically had a lower priority in this field. Until
now, research into BCI technology has usually focused on assistive applications, such as spelling
devices, wheelchair control or neuroprostheses rather than applications with entertainment purposes.
However, interest in entertainment applications has arisen over the recent years due to the significant
advances in this technology. In fact, improvements in its performance have opened the way to
extending BCI use to non-disabled people. BCIs create a new interaction modality which may turn
video games into even more challenging and attractive experiences. Additionally, BCI may provide a
way of accessing knowledge on the user’s experiences, thereby improving games through information
from brain activity. BCIs can report when the gamer is bored, anxious or frustrated with the aim of
using this knowledge for designing future games [295].
Figure 10. Pacman game. The gamer has to move through the maze to reach the exit in the
right wall. The shortest path is marked with gray track marks, but the gamer can decide to
run the rest of maze to receive additional credits (adapted from [296]).
Nose:
Direction
of next
move.
Discrimination Line:
Indication of competing
laterality intentions
S
Sensors 2012, 12 12556
Figure 111. (a) Emotivv EPOC neuuroheadset [5]; (b) Neuurosky Minndwave [6].
(a) (b))
neuroheadset (Figure 11(b)) with software applications that can respond to user brainwaves or mental
states. Likewise, it provides a set of software tools for developers. Also, large software companies such
as Microsoft have shown interest in BCI research, exploring the development of pilot novel
applications that use BCIs [301].
BCI systems have also been used in a broad variety of applications beyond the traditional areas of
communication, motor restoration, environmental control, locomotion, and entertainment. The ability
of BCI feedback to induce cortical plasticity may be the basis for medical applications. Users can
acquire selective control over certain brain areas by means of neurofeedback, with the aim of inducing
behavioral changes in the brain. Neurofeedback provided by a BCI system may improve cognitive
performance [302,303], speech skills [304], affection [305], and pain management [306], and has been
used in the treatment of mental disorders, such as epilepsy [307,308], attention deficit [309],
schizophrenia [310], depression [311], alcohol dependence [312], or paedophilia [313]. On the other
hand, brain signal recordings can be used in an assessment of brain functions to evaluate their status in
health and disease [314].
The opportunity to examine brain signals can also be commercially exploited. Neuromarketing is a
relatively young field of research that applies neuroscientific methods to marketing research. To date,
few neuromarketing studies have been conducted, although some evidence has been found to suggest
that neuroimaging could have a role in several areas of marketing [315–318]. Neuromarketing may
provide a more efficient trade-off between costs and benefits. Product concepts could be tested by
means of neuromarketing, removing those that are not promising at the start of the manufacturing
process. This would lead to a more efficient distribution of sources, because only the more promising
products would be developed [319]. In addition, neuromarketing may be a source of more accurate
information on the underlying preferences of the users, rather than data from standard market research
studies [319]. Neuroimaging may reveal hidden information on consumers’ true preferences that
cannot be explicitly expressed. The brain’s response to advertisements could be measured and the
effectiveness of advertising campaigns could therefore be quantified.
Despite it being an emerging field, several companies such as Neurofocus [320], Neuroconsult [321],
Neuro Insight [322] or EmSense [323], among others, currently offer neuromarketing services. It is
also attracting increasing attention among researchers. The field has raised some ethical issues
concerning this technology, in as much as it may be able to manipulate the brain and consumer
behavior [324].
9. Conclusions
This article has reviewed the state-of-the-art of BCI systems, discussing fundamental aspects of BCI
system design. The most significant goals that have driven BCI research over the last 20 years have
been presented. It has been noted that many breakthroughs were achieved in BCI research. Different
neuroimaging approaches have been successfully applied in BCI: (i) EEG, which provides acceptable
quality signals with high portability and is by far the most usual modality in BCI; (ii) fMRI and MEG,
which are proven and effective methods for localizing active regions inside the brain; (iii) NIRS,
Sensors 2012, 12 1258
which is a very promising neuroimaging method in BCI; and (iv) invasive modalities, which have been
presented as valuable methods to provide the high quality signals required in some multidimensional
control applications e.g., neuroprostheses control.
A wide variety of signal features and classification algorithms have been tested in the BCI design.
Although BCI research is relatively young, many advances have been achieved in a little over two
decades, because many of these methods are based on previous signal processing and pattern
recognition research. Many studies have demonstrated the valuable accuracy of BCIs and provided
acceptable information bit rate, despite the inherent major difficulties in brain signal processing.
Accordingly, user training time has been significantly reduced, which has led to more widespread BCI
applications in the daily life of disabled people, such as word processing, browsers, email, wheelchair
control, simple environmental control or neuroprostheses among others.
In spite of the recent important advances in the BCI field, some issues still need to be solved. First,
the relative advantages and disadvantages of the different signal acquisition methods are still unclear.
Their clarification will require further human and animal studies. Second, invasive methods need
further investigation to deal with tissue damage, risk of infection, and long-term stability concerns.
Electrodes that contain neurotropic mediums that promote neuronal growth and wireless transmission
of neuronal signals recorded have already been proposed. Third, the electrophysiological and
metabolic signals that are best able to encode user intent should be better identified and characterized.
The majority of BCI studies have treated time, frequency, and spatial dimensions of brain signals
independently. These signal dimension interdependencies may lead to significant improvement in BCI
performance. Fourth, information bit rate provided by current BCIs is low for effective human-
machine interaction in some applications. Exogenous-based BCI may provide much higher throughput.
Fifth, the unsupervised adaptation is a key challenge for BCI deployment outside the lab. Some
moderately successful adaptive classification algorithms have already been proposed. And finally,
most BCI applications are at the research stage and they are not ready to be introduced into people’s
homes for continuous use in their daily life. In addition to their low information transfer rates and
variable reliability, most current BCI systems are uncomfortable, because the electrodes need to be
moistened, the software may require initiation, and the electrode contacts need continuous correction.
An easy-to-use P300-based BCI with remote monitoring using a high-speed internet connection has
already been proposed to reduce dependence on technical experts.
The latest advances in BCI research suggest that innovative developments may be forthcoming in
the near future. These achievements and the potential for new BCI applications have obviously given a
significant boost to BCI research involving multidisciplinary scientists e.g., neuroscientists, engineers,
mathematicians, and clinical rehabilitation specialists, among others. Interest in the BCI field is
expected to increase and BCI design and development will in all probability continue to bring benefits
to the daily lives of disabled people. Furthermore, recent commercial interest within certain companies
suggests that BCI systems may find useful applications in the general population, and not just for
people living with severe disabilities. In the near future, BCI systems may therefore become a new
mode of human-machine interaction with levels of everyday use that are similar to other current interfaces.
Sensors 2012, 12 1259
Acknowledgements
This work was partially supported by the regional 2010 Research Project Plan of the Junta de
Castilla y León, (Spain), under project VA034A10-2.
References
1. Khalid, M.B.; Rao, N.I.; Rizwan-i-Haque, I.; Munir, S.; Tahir, F. Towards a Brain Computer
Interface Using Wavelet Transform with Averaged and Time Segmented Adapted Wavelets. In
Proceedings of the 2nd International Conference on Computer, Control and Communication
(IC4’09) Karachi, Sindh, Pakistan, February 2009; pp. 1–4.
2. Wolpaw, J.R.; Birbaumer, N.; McFarland, D.J.; Pfurtscheller, G.; Vaughan, T.M.
Brain-computer interfaces for communication and control. Clin. Neurophysiol. 2002, 113,
767–791.
3. Wolpaw, J.R. Brain–computer interfaces as new brain output pathways. J. Physiol. 2007, 579,
613–619.
4. Konrad, P.; Shanks, T. Implantable brain computer interface: Challenges to neurotechnology
translation. Neurobiol. Dis. 2010, 38, 369–375.
5. Emotiv—Brain Computer Interface Technology. Available online: http://www.emotiv.com
(accessed on 12 July 2011).
6. NeuroSky—Brainwave Sensors for Everybody. Available online: http://www.neurosky.com
(accessed on 12 July 2011).
7. Blankertz, B.; Tangermann, M.; Vidaurre, C.; Fazli, S.; Sannelli, C.; Haufe, S.; Maeder, C.;
Ramsey, L.; Sturm, I.; Curio, G.; Mueller, K.R. The Berlin Brain-Computer Interface:
Non-Medical Uses of BCI Technology. Front. Neurosci. 2010, doi: 10.3389/fnins.2010.00198.
8. Wolpaw, J.R.; Birbaumer, N.; Heetderks, W.J.; McFarland, D.J.; Peckham, P.H.; Schalk, G.;
Donchin, E.; Quatrano, L.A.; Robinson, C.J.; Vaughan, T.M. Brain-computer interface
technology: A review of the first international meeting. IEEE Trans. Rehabil. Eng. 2000, 8,
164–173.
9. Mason, S.G.; Birch, G.E. A general framework for brain-computer interface design. IEEE Trans.
Neural Syst. Rehabil. Eng. 2003, 11, 70–85.
10. Baillet, S.; Mosher, J.C.; Leahy, R.M. Electromagnetic brain mapping. IEEE Signal Process.
Mag. 2001, 18, 14–30.
11. Laureys, S.; Boly, M.; Tononi, G.; Functional Neuroimaging. In The Neurology of Consciousness;
Steven, L., Giulio, T., Eds.; Academic Press: New York, NY, USA, 2009; pp. 31–42.
12. Sellers, E.W.; Vaughan, T.M.; Wolpaw, J.R. A brain-computer interface for long-term
independent home use. Amyotroph. Lateral Scler. 2010, 11, 449–455.
13. Muller-Putz, G.R.; Pfurtscheller, G. Control of an Electrical Prosthesis With an SSVEP-Based
BCI. IEEE Trans. Biomed. Eng. 2008, 55, 361–364.
14. Cincotti, F.; Mattia, D.; Aloise, F.; Bufalari, S.; Schalk, G.; Oriolo, G.; Cherubini, A.;
Marciani, M.G.; Babiloni, F. Non-invasive brain-computer interface system: Towards its
application as assistive technology. Brain Res. Bull. 2008, 75, 796–803.
Sensors 2012, 12 1260
15. Lebedev, M.A.; Nicolelis, M.A.L. Brain-machine interfaces: Past, present and future. Trends
Neurosci. 2006, 29, 536–546.
16. Kennedy, P.R.; Bakay, R.A.E.; Moore, M.M.; Adams, K.; Goldwaithe, J. Direct control of a
computer from the human central nervous system. IEEE Trans. Rehabil. Eng. 2000, 8, 198–202.
17. Bossetti, C.A.; Carmena, J.M.; Nicolelis, M.A.L.; Wolf, P.D. Transmission latencies in a
telemetry-linked brain-machine interface. IEEE Trans. Biomed. Eng. 2004, 51, 919–924.
18. Teplan, M. Fundamentals of EEG measurement. Meas. Sci. Rev. 2002, 2, 1–11.
19. Sinclair, C.M.; Gasper, M.C.; Blum, A.S.; Basic Electronics in Clinical Neurophysiology. In The
Clinical Neurophysiology Primer; Blum, A.S., Rutkove, S.B., Eds.; Humana Press Inc.: Totowa,
NJ, USA, 2007; pp. 3–18.
20. Usakli, A.B. Improvement of EEG signal acquisition: An electrical aspect for state of the art of
front end. Comput. Intell. Neurosci. 2010, 2010, 630649.
21. Fonseca, C.; Cunha, J.P.S.; Martins, R.E.; Ferreira, V.M.; de Sa, J.P.M.; Barbosa, M.A.;
da Silva, A.M. A Novel Dry Active Electrode for EEG Recording. IEEE Trans. Biomed. Eng.
2007, 54, 162–165.
22. Taheri, B.A.; Knight, R.T.; Smith, R.L. A dry electrode for EEG recording. Electroencephalogr.
Clin. Neurophysiol. 1994, 90, 376–383.
23. Gargiulo, G.; Calvo, R.A.; Bifulco, P.; Cesarelli, M.; Jin, C.; Mohamed, A.; van Schaik, A. A
new EEG recording system for passive dry electrodes. Clin. Neurophysiol. 2010, 121, 686–693.
24. Leach, W.M., Jr. Fundamentals of low-noise analog circuit design. Proc. IEEE 1994, 82, 1515–1538.
25. Kübler, A.; Kotchoubey, B.; Kaiser, J.; Wolpaw, J.R.; Birbaumer, N. Brain-Computer
Communication: Unlocking the Locked in; American Psychological Association: Washington,
DC, USA, 2001.
26. Anand, B.K.; Chhina, G.S.; Singh, B. Some aspects of electroencephalographic studies in Yogis.
Electroencephalogr. Clin. Neurophysiol. 1961, 13, 452–456.
27. Aftanas, L.I.; Golocheikine, S.A. Human anterior and frontal midline theta and lower alpha
reflect emotionally positive state and internalized attention: High-resolution EEG investigation of
meditation. Neurosci. Lett. 2001, 310, 57–60.
28. Fernández, T.; Harmony, T.; Rodríguez, M.; Bernal, J.; Silva, J.; Reyes, A.; Marosi, E. EEG
activation patterns during the performance of tasks involving different components of mental
calculation. Electroencephalogr. Clin. Neurophysiol. 1995, 94, 175–182.
29. Caplan, J.B.; Madsen, J.R.; Raghavachari, S.; Kahana, M.J. Distinct Patterns of Brain
Oscillations Underlie Two Basic Parameters of Human Maze Learning. J. Neurophysiol. 2001,
86, 368–380.
30. Klimesch, W.; Doppelmayr, M.; Yonelinas, A.; Kroll, N.E.A.; Lazzara, M.; Röhm, D.;
Gruber, W. Theta synchronization during episodic retrieval: Neural correlates of conscious
awareness. Cogn. Brain. Res. 2001, 12, 33–38.
31. Jaime A.P. The functional significance of mu rhythms: Translating “seeing” and “hearing” into
“doing”. Brain Res. Rev. 2005, 50, 57–68.
32. Black, A.H. The Operant Conditioning of Central Nervous System Electrical Activity. In
Psychology of Learning and Motivation, Gordon, H.B., Ed.; Academic Press: New York, NY,
USA, 1972; Volume 6, pp. 47–95.
Sensors 2012, 12 1261
33. Klimesch, W. EEG-alpha rhythms and memory processes. Int. J. Psychophysiol. 1997, 26, 319–340.
34. Venables, L.; Fairclough, S. The influence of performance feedback on goal-setting and mental
effort regulation. Motiv. Emotion 2009, 33, 63–74.
35. Pfurtscheller, G.; Brunner, C.; Schlögl, A.; Lopes da Silva, F.H. Mu rhythm (de)synchronization and
EEG single-trial classification of different motor imagery tasks. Neuroimage 2006, 31, 153–159.
36. Pfurtscheller, G.; Neuper, C. Motor imagery and direct brain-computer communication. Proc.
IEEE 2001, 89, 1123–1134.
37. Lee, K.-H.; Williams, L.M.; Breakspear, M.; Gordon, E. Synchronous Gamma activity: A review
and contribution to an integrative neuroscience model of schizophrenia. Brain Res. Rev. 2003,
41, 57–78.
38. Brown, P.; Salenius, S.; Rothwell, J.C.; Hari, R. Cortical correlate of the piper rhythm in
humans. J. Neurophysiol. 1998, 80, 2911–2917.
39. Mima, T.; Simpkins, N.; Oluwatimilehin, T.; Hallett, M. Force level modulates human cortical
oscillatory activities. Neurosci. Lett. 1999, 275, 77–80.
40. Lutzenberger, W.; Pulvermüller, F.; Elbert, T.; Birbaumer, N. Visual stimulation alters local
40-Hz responses in humans: an EEG-study. Neurosci. Lett. 1995, 183, 39–42.
41. Müller, M.M.; Keil, A.; Gruber, T.; Elbert, T. Processing of affective pictures modulates
right-hemispheric gamma band EEG activity. Clin. Neurophysiol. 1999, 110, 1913–1920.
42. Müller, M.M.; Bosch, J.; Elbert, T.; Kreiter, A.; Sosa, M.V.; Sosa, P.V.; Rockstroh, B. Visually
induced gamma-band responses in human electroencephalographic activity—a link to animal
studies. Exp. Brain Res. 1996, 112, 96–102.
43. Zhang, L.; He, W.; He, C.; Wang, P. Improving Mental task classification by adding high
frequency band information. J. Med. Syst. 2010, 34, 51–60.
44. Darvas, F.; Scherer, R.; Ojemann, J.G.; Rao, R.P.; Miller, K.J.; Sorensen, L.B. High gamma
mapping using EEG. Neuroimage 2010, 49, 930–938.
45. Miller, K.J.; Leuthardt, E.C.; Schalk, G.; Rao, R.P.N.; Anderson, N.R.; Moran, D.W.;
Miller, J.W.; Ojemann, J.G. Spectral changes in cortical surface potentials during motor
movement. J. Neurosci. 2007, 27, 2424–2432.
46. Jasper, H.H. The ten-twenty electrode system of the International Federation. Electroencephalogr.
Clin. Neurophysiol. 1958, 10, 371–375.
47. Waldert, S.; Pistohl, T.; Braun, C.; Ball, T.; Aertsen, A.; Mehring, C. A review on directional
information in neural signals for brain-machine interfaces. J. Phisiol. Paris 2009, 103, 244–254.
48. Salmelin, R.; Hámáaláinen, M.; Kajola, M.; Hari, R. Functional segregation of movement-related
rhythmic activity in the human brain. Neuroimage 1995, 2, 237–243.
49. Babiloni, C.; Pizzella, V.; Gratta, C.D.; Ferretti, A.; Romani, G.L.; Fundamentals of
Electroencefalography, Magnetoencefalography, and Functional Magnetic Resonance Imaging.
In Brain Machine Interfaces for Space Applications; Luca, R., Dario, I., Leopold, S., Eds.;
Academic Press: New York, NY, USA, 2009; Volume 86, pp. 67–80.
50. Mellinger, J.; Schalk, G.; Braun, C.; Preissl, H.; Rosenstiel, W.; Birbaumer, N.; Kübler, A. An
MEG-based brain-computer interface (BCI). NeuroImage 2007, 36, 581–593.
Sensors 2012, 12 1262
51. Wang, W.; Sudre, G.P.; Xu, Y.; Kass, R.E.; Collinger, J.L.; Degenhart, A.D.; Bagic, A.I.;
Weber, D.J. Decoding and cortical source localization for intended movement direction with
MEG. J. Neurophysiol. 2010, 104, 2451–2461.
52. Lal, T.N.; Schröder, M.; Hill, N.J.; Preissl, H.; Hinterberger, T.; Mellinger, J.; Bogdan, M.;
Rosenstiel, W.; Hofmann, T.; Birbaumer, N.; Schölkopf, B. A Brain Computer Interface with
Online Feedback Based on Magnetoencephalography. In Proceedings of the 22nd International
Conference on Machine Learning (ICML’ 05), Bonn, Germany, 7–11 August 2005; pp. 465–472.
53. Kauhanen, L.; Nykopp, T.; Lehtonen, J.; Jylanki, P.; Heikkonen, J.; Rantanen, P.; Alaranta, H.;
Sams, M. EEG and MEG brain-computer interface for tetraplegic patients. IEEE Trans. Neural
Syst. Rehabil. Eng. 2006, 14, 190–193.
54. Georgopoulos, A.; Langheim, F.; Leuthold, A.; Merkle, A. Magnetoencephalographic signals
predict movement trajectory in space. Exp. Brain Res. 2005, 167, 132–135.
55. Mellinger, J.; Schalk, G.; Braun, C.; Preissl, H.; Rosenstiel, W.; Birbaumer, N.; Kübler, A.
An MEG-based brain-computer interface (BCI). Neuroimage 2007, 36, 581–593.
56. Jinyin, Z.; Sudre, G.; Xin, L.; Wei, W.; Weber, D.J.; Bagic, A. Clustering linear discriminant
analysis for MEG-Based brain computer interfaces. IEEE Trans. Neural Syst. Rehabil. Eng.
2011, 19, 221–231.
57. Sabra, N.I.; Wahed, M.A. The Use of MEG-Based Brain Computer Interface for Classification of
Wrist Movements in Four Different Directions. In Proceedings of the 28th National Radio
Science Conference (NRSC’ 11), Cairo, Egypt, April 2011; pp. 1–7.
58. Ball, T.; Kern, M.; Mutschler, I.; Aertsen, A.; Schulze-Bonhage, A. Signal quality of
simultaneously recorded invasive and non-invasive EEG. Neuroimage 2009, 46, 708–716.
59. Loeb, G.E.; Walker, A.E.; Uematsu, S.; Konigsmark, B.W. Histological reaction to various
conductive and dielectric films chronically implanted in the subdural space. J. Biomed. Mater.
Res. 1977, 11, 195–210.
60. Bullara, L.A.; Agnew, W.F.; Yuen, T.G.H.; Jacques, S.; Pudenz, R.H. Evaluation of electrode
array material for neural prostheses. Neurosurgery 1979, 5, 681–686.
61. Yuen, T.G.H.; Agnew, W.F.; Bullara, L.A. Tissue response to potential neuroprosthetic materials
implanted subdurally. Biomaterials 1987, 8, 138–141.
62. Margalit, E.; Weiland, J.D.; Clatterbuck, R.E.; Fujii, G.Y.; Maia, M.; Tameesh, M.; Torres, G.;
D'Anna, S.A.; Desai, S.; Piyathaisere, D.V.; et al. Visual and electrical evoked response recorded
from subdural electrodes implanted above the visual cortex in normal dogs under two methods of
anesthesia. J. Neurosci. Methods 2003, 123, 129–137.
63. Chao, Z.C.; Nagasaka, Y.; Fujii, N. Long-term asynchronous decoding of arm motion using
electrocorticographic signals in monkeys. Front. Neuroeng. 2010, 3, doi: 10.3389/fneng.2010.00003.
64. Matsuo, T.; Kawasaki, K.; Osada, T.; Sawahata, H.; Suzuki, T.; Shibata, M.; Miyakawa, N.;
Nakahara, K.; Iijima, A.; Sato, N.; et al. Intrasulcal electrocorticography in macaque monkeys
with minimally invasive neurosurgical protocols. Front. Syst. Neurosci. 2011, doi: 10.3389/
fnsys.2011.00034.
65. Crone, N.E.; Miglioretti, D.L.; Gordon, B.; Sieracki, J.M.; Wilson, M.T.; Uematsu, S.; Lesser, R.P.
Functional mapping of human sensorimotor cortex with electrocorticographic spectral analysis. I.
Alpha and beta event-related desynchronization. Brain 1998, 121, 2271–2299.
Sensors 2012, 12 1263
66. Crone, N.E.; Miglioretti, D.L.; Gordon, B.; Lesser, R.P. Functional mapping of human sensorimotor
cortex with electrocorticographic spectral analysis. II. Event-related synchronization in the gamma
band. Brain 1998, 121, 2301–2315.
67. Miller, K.J.; denNijs, M.; Shenoy, P.; Miller, J.W.; Rao, R.P.N.; Ojemann, J.G. Real-time
functional brain mapping using electrocorticography. NeuroImage 2007, 37, 504–507.
68. Levine, S.P.; Huggins, J.E.; BeMent, S.L.; Kushwaha, R.K.; Schuh, L.A.; Passaro, E.A.;
Rohde, M.M.; Ross, D.A. Identification of electrocorticogram patterns as the basis for a direct
brain interface. J. Clin. Neurophysiol. 1999, 16, 439–447.
69. Leuthardt, E.C.; Schalk, G.; Wolpaw, J.R.; Ojemann, J.G; Moran, D.W. A brain-computer
interface using electrocorticographic signals in humans. J. Neural Eng. 2004, doi: 10.1088/1741-
2560/1/2/001.
70. Schalk, G.; Schalk, G.; Kubánek, J.; Miller, K.J.; Anderson, N.R.; Leuthardt, E.C.; Ojemann, J.G.;
Limbrick, D.; Moran, D.; Gerhardt, L.A.; et al. Decoding two-dimensional movement trajectories
using electrocorticographic signals in humans. J. Neural Eng. 2007, doi: 10.1088/1741-2560/4/3/012.
71. Polikov, V.S.; Tresco, P.A.; Reichert, W.M. Response of brain tissue to chronically implanted
neural electrodes. J. Neurosci. Methods 2005, 148, 1–18.
72. Maynard, E.M.; Nordhausen, C.T.; Normann, R.A. The Utah Intracortical Electrode Array:
A recording structure for potential brain-computer interfaces. Electroencephalogr. Clin.
Neurophysiol. 1997, 102, 228–239.
73. Lauer, R.T.; Peckham, P.H.; Kilgore, K.L.; Heetderks, W.J. Applications of cortical signals to
neuroprosthetic control: A critical review. IEEE Trans. Rehabil. Eng. 2000, 8, 205–208.
74. Georgopoulos, A.; Schwartz, A.; Kettner, R. Neuronal population coding of movement direction.
Science 1986, 233, 1416–1419.
75. Schwartz, A.B. Motor cortical activity during drawing movements: population representation
during sinusoid tracing. J. Neurophysiol. 1993, 70, 28–36.
76. Wessberg, J.; Stambaugh, C.R.; Kralik, J.D.; Beck, P.D.; Laubach, M.; Chapin, J.K.; Kim, J.;
Biggs, S.J.; Srinivasan, M.A.; Nicolelis, M.A.L. Real-time prediction of hand trajectory by
ensembles of cortical neurons in primates. Nature 2000, 408, 361–365.
77. Taylor, D.M.; Tillery, S.I.H.; Schwartz, A.B. Direct cortical control of 3D neuroprosthetic
devices. Science 2002, 296, 1829–1832.
78. Jackson, A.; Mavoori, J.; Fetz, E.E. Correlations between the same motor cortex cells and arm
muscles during a trained task, free behavior, and natural sleep in the macaque monkey.
J. Neurophysiol. 2007, 97, 360–374.
79. Velliste, M.; Perel, S.; Spalding, M.C.; Whitford, A.S.; Schwartz, A.B. Cortical control of a
prosthetic arm for self-feeding. Nature 2008, 453, 1098–1101.
80. Vargas-Irwin, C.E.; Shakhnarovich, G.; Yadollahpour, P.; Mislow, J.M.K.; Black, M.J.;
Donoghue, J.P. Decoding complete reach and grasp actions from local primary motor cortex
populations. J. Neurosci. 2010, 30, 9659–9669.
81. Carpaneto, J.; Umiltà, M.A.; Fogassi, L.; Murata, A.; Gallese, V.; Micera, S.; Raos, V. Decoding
the activity of grasping neurons recorded from the ventral premotor area F5 of the macaque
monkey. Neuroscience 2011, 188, 80–94.
Sensors 2012, 12 1264
82. Kennedy, P.R.; Kirby, M.T.; Moore, M.M.; King, B.; Mallory, A. Computer control using human
intracortical local field potentials. IEEE Trans. Neural Syst. Rehabil. Eng. 2004, 12, 339–344.
83. deCharms, R.C.; Christoff, K.; Glover, G.H.; Pauly, J.M.; Whitfield, S.; Gabrieli, J.D.E. Learned
regulation of spatially localized brain activation using real-time fMRI. Neuroimage 2004, 21,
436–443.
84. Weiskopf, N.; Mathiak, K.; Bock, S.W.; Scharnowski, F.; Veit, R.; Grodd, W.; Goebel, R.;
Birbaumer, N. Principles of a brain-computer interface (BCI) based on real-time functional
magnetic resonance imaging (fMRI). IEEE Trans. Biomed. Eng. 2004, 51, 966–970.
85. Lee, J.; Ryu, J.; Jolesz, F.A.; Cho, Z.H.; Yoo, S.S. Brain-machine interface via real-time fMRI:
Preliminary study on thought-controlled robotic arm. Neurosci. Lett. 2009, 450, 1–6.
86. Logothetis, N.K.; Pauls, J.; Augath, M.; Trinath, T.; Oeltermann, A. Neurophysiological
investigation of the basis of the fMRI signal. Nature 2001, 412, 150–157.
87. Weiskopf, N.; Sitaram, R.; Josephs, O.; Veit, R.; Scharnowski, F.; Goebel, R.; Birbaumer, N.;
Deichmann, R.; Mathiak, K. Real-time functional magnetic resonance imaging: Methods and
applications. Magn. Reson. Imaging 2007, 25, 989–1003.
88. Christopher deCharms, R. Applications of real-time fMRI. Nat. Rev. Neurosci. 2008, 9, 720–729.
89. Moench, T.; Hollmann, M.; Grzeschik, R.; Mueller, C.; Luetzkendorf, R.; Baecke, S.;
Luchtmann, M.; Wagegg, D.; Bernarding, J. Real-Time Classification of Activated Brain Areas
for fMRI-Based Human-Brain-Interfaces; Society of Photo-Optical Instrumentation Engineers:
Belligham, WA, USA, 2008.
90. Ward, B.D.; Mazaheri, Y. Information transfer rate in fMRI experiments measured using mutual
information theory. J. Neurosci. Methods 2008, 167, 22–30.
91. Coyle, S.M.; Ward, T.E.; Markham, C.M. Brain-computer interface using a simplified functional
near-infrared spectroscopy system. J. Neural Eng. 2007, doi: 10.1088/1741-2560/4/3/007.
92. Taga, G.; Homae, F.; Watanabe, H. Effects of source-detector distance of near infrared
spectroscopy on the measurement of the cortical hemodynamic response in infants. NeuroImage
2007, 38, 452–460.
93. Kennan, R.P.; Horovitz, S.G.; Maki, A.; Yamashita, Y.; Koizumi, H.; Gore, J.C. Simultaneous
recording of event-related auditory oddball response using transcranial near infrared optical
topography and surface EEG. Neuroimage 2002, 16, 587–592.
94. Izzetoglu, M.; Devaraj, A.; Bunce, S.; Onaral, B. Motion artifact cancellation in NIR
spectroscopy using Wiener filtering. IEEE Trans. Biomed. Eng. 2005, 52, 934–938.
95. Strangman, G.; Boas, D.A.; Sutton, J.P. Non-invasive neuroimaging using near-infrared light.
Biol. Psychiatry 2002, 52, 679–693.
96. Cui, X.; Bray, S.; Reiss, A.L. Functional near infrared spectroscopy (NIRS) signal improvement
based on negative correlation between oxygenated and deoxygenated hemoglobin dynamics.
Neuroimage 2010, 49, 3039–3046.
97. Shirley, C.; Ward, T.; Markham, C.; McDarby, G. On the suitability of near-infrared (NIR)
systems for next-generation brain-computer interfaces. Physiol. Meas. 2004, 25, 815.
98. Power, S.D.; Kushki, A.; Chau, T. Towards a system-paced near-infrared spectroscopy
brain-computer interface: Differentiating prefrontal activity due to mental arithmetic and mental
singing from the no-control state. J. Neural Eng. 2011, 8, 066004.
Sensors 2012, 12 1265
99. Villringer, A.; Planck, J.; Hock, C.; Schleinkofer, L.; Dirnagl, U. Near infrared spectroscopy
(NIRS): A new tool to study hemodynamic changes during activation of brain function in human
adults. Neurosci. Lett. 1993, 154, 101–104.
100. Sitaram, R.; Zhang, H.; Guan, C.; Thulasidas, M.; Hoshi, Y.; Ishikawa, A.; Shimizu, K.;
Birbaumer, N. Temporal classification of multichannel near-infrared spectroscopy signals of
motor imagery for developing a brain–computer interface. Neuroimage 2007, 34, 1416–1427.
101. Regan, D. Human Brain Electrophysiology: Evoked Potentials and Evoked Magnetic Fields in
Science and Medicine; Elsevier: New York, NY, USA, 1989.
102. Yijun, W.; Ruiping, W.; Xiaorong, G.; Bo, H.; Shangkai, G. A practical VEP-based
brain-computer interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 234–240.
103. Jinghai, Y.; Derong, J.; Jianfeng, H. Design and Application of Brain-Computer Interface Web
Browser Based on VEP. In Proceedings of the International Conference on Future BioMedical
Information Engineering (FBIE’09), Sanya, China, 13–14 December 2009; pp. 77–80.
104. Xiaorong, G.; Dingfeng, X.; Ming, C.; Shangkai, G. A BCI-based environmental controller for
the motion-disabled. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 137–140.
105. Odom, J.V.; Bach, M.; Barber, C.; Brigell, M.; Marmor, M.F.; Tormene, A.P.; Holder, G.E.
Visual evoked potentials standard (2004). Doc. Ophthalmol. 2004, 108, 115–123.
106. Zhu, D.; Bieger, J.; Garcia Molina, G.; Aarts, R.M. A survey of stimulation methods used in
SSVEP-Based BCIs. Comput. Intell. Neurosci. 2010, doi: 10.1155/2010/702357.
107. Galloway, N. Human brain electrophysiology: Evoked potentials and evoked magnetic fields in
science and medicine. Br. J. Ophthalmol. 1990, 74, 255.
108. Perlstein, W.M.; Cole, M.A.; Larson, M.; Kelly, K.; Seignourel, P.; Keil, A. Steady-state visual
evoked potentials reveal frontally-mediated working memory activity in humans. Neurosci. Lett.
2003, 342, 191–195.
109. Gray, M.; Kemp, A.H.; Silberstein, R.B.; Nathan, P.J. Cortical neurophysiology of anticipatory
anxiety: An investigation utilizing steady state probe topography (SSPT). Neuroimage 2003, 20,
975–986.
110. Guangyu, B.; Xiaorong, G.; Yijun, W.; Bo, H.; Shangkai, G. VEP-based brain-computer
interfaces: Time, frequency, and code modulations [Research Frontier]. IEEE Comput. Intell.
Mag. 2009, 4, 22–26.
111. Lee, P.; Hsieh, J.; Wu, C.; Shyu, K.; Wu, Y. Brain computer interface using flash onset and
offset visual evoked potentials. Clin. Neurophysiol. 2008, 119, 605–616.
112. Allison, B.Z.; McFarland, D.J.; Schalk, G.; Zheng, S.D.; Jackson, M.M.; Wolpaw, J.R. Towards
an independent brain-computer interface using steady state visual evoked potentials. Clin.
Neurophysiol. 2008, 119, 399–408.
113. Dan, Z.; Alexander, M.; Xiaorong, G.; Bo, H.; Andreas, K.E.; Shangkai, G. An independent
brain-computer interface using covert non-spatial visual selective attention. J. Neural Eng. 2010,
7, 016010.
114. Yijun, W.; Xiaorong, G.; Bo, H.; Chuan, J.; Shangkai, G. Brain-computer interfaces based on
visual evoked potentials. IEEE Eng. Med. Biol. Mag. 2008, 27, 64–71.
115. Wu, Z.; Lai, Y.; Xia, Y.; Wu, D.; Yao, D. Stimulator selection in SSVEP-based BCI. Med. Eng.
Phys. 2008, 30, 1079–1088.
Sensors 2012, 12 1266
116. Birbaumer, N.; Elbert, T.; Canavan, A.G.; Rockstroh, B. Slow potentials of the cerebral cortex
and behavior. Physiol. Rev. 1990, 70, 1–41.
117. Hinterberger, T.; Schmidt, S.; Neumann, N.; Mellinger, J.; Blankertz, B.; Curio, G.; Birbaumer, N.
Brain-computer communication and slow cortical potentials. IEEE Trans. Biomed. Eng. 2004,
51, 1011–1018.
118. Kaiser, J. Self-initiation of EEG-based communication in paralyzed patients. Clin. Neurophysiol.
2001, 112, 551–554.
119. Neumann, N.; Birbaumer, N. Predictors of successful self control during brain-computer
communication. J. Neurol. Neurosurg. Psychiatry 2003, 74, 1117–1121.
120. Kübler, A.; Kotchoubey, B.; Hinterberger, T.; Ghanayim, N.; Perelmouter, J.; Schauer, M.;
Fritsch, C.; Taub, E.; Birbaumer, N. The thought translation device: A neurophysiological
approach to communication in total motor paralysis. Exp. Brain Res. 1999, 124, 223–232.
121. Birbaumer, N.; Hinterberger, T.; Kubler, A.; Neumann, N. The thought-translation device (TTD):
Neurobehavioral mechanisms and clinical outcome. IEEE Trans. Neural Syst. Rehabil. Eng.
2003, 11, 120–123.
122. Iversen, I.H.; Ghanayim, N.; Kübler, A.; Neumann, N.; Birbaumer, N.; Kaiser, J. A
brain-computer interface tool to assess cognitive functions in completely paralyzed patients with
amyotrophic lateral sclerosis. Clin. Neurophysiol. 2008, 119, 2214–2223.
123. Farwell, L.A.; Donchin, E. Talking off the top of your head: Toward a mental prosthesis utilizing
event-related brain potentials. Electroencephalogr. Clin. Neurophysiol. 1988, 70, 510–523.
124. Donchin, E.; Smith, D.B.D. The contingent negative variation and the late positive wave of the
average evoked potential. Electroencephalogr. Clin. Neurophysiol. 1970, 29, 201–203.
125. Polich, J.; Ellerson, P.C.; Cohen, J. P300, stimulus intensity, modality, and probability. Int. J.
Psychophysiol. 1996, 23, 55–62.
126. Ravden, D.; Polich, J. On P300 measurement stability: Habituation, intra-trial block variation,
and ultradian rhythms. Biol. Psychol. 1999, 51, 59–76.
127. Tanaka, K.; Matsunaga, K.; Wang, H.O. Electroencephalogram-based control of an electric
wheelchair. IEEE Trans. Robot. 2005, 21, 762–766.
128. Mugler, E.M.; Ruf, C.A.; Halder, S.; Bensch, M.; Kubler, A. Design and implementation of a
P300-based brain-computer interface for controlling an internet browser. IEEE Trans. Neural
Syst. Rehabil. Eng. 2010, 18, 599–609.
129. Furdea, A.; Halder, S.; Krusienski, D.J.; Bross, D.; Nijboer, F.; Birbaumer, N.; Kübler, A. An
auditory oddball (P300) spelling system for brain-computer interfaces. Psychophysiology 2009,
46, 617–625.
130. Rivet, B.; Souloumiac, A.; Attina, V.; Gibert, G. xDAWN algorithm to enhance evoked potentials:
Application to brain-computer interface. IEEE Trans. Biomed. Eng. 2009, 56, 2035–2043.
131. Rakotomamonjy, A.; Guigue, V. BCI Competition III: Dataset II-Ensemble of SVMs for BCI
P300 Speller. IEEE Trans. Biomed. Eng. 2008, 55, 1147–1154.
132. Salvaris, M.; Sepulveda, F. Visual modifications on the P300 speller BCI paradigm. J. Neural
Eng. 2009, 6, 046011.
Sensors 2012, 12 1267
133. Takano, K.; Komatsu, T.; Hata, N.; Nakajima, Y.; Kansaku, K. Visual stimuli for the P300
brain-computer interface: A comparison of white/gray and green/blue flicker matrices. Clin.
Neurophysiol. 2009, 120, 1562–1566.
134. Ikegami, S.; Takano, K.; Saeki, N.; Kansaku, K. Operation of a P300-based brain-computer
interface by individuals with cervical spinal cord injury. Clin. Neurophysiol. 2011, 122, 991–996.
135. Hill, J.; Farquhar, J.; Martens, S.; Bießmann, F.; Schölkopf, B. Effects of stimulus type and of
error-correcting code design on BCI speller performance. In Proceedings of the Twenty-Second
Annual Conference on Neural Information Processing Systems (NIPS 2008), Vancouver,
Canada, 8–11 December 2008.
136. Martens, S.M.M.; Hill, N.J.; Farquhar, J.; Schölkopf, B. Overlap and refractory effects in a
brain-computer interface speller based on the visual P300 event-related potential. J. Neural Eng.
2009, 6, 026003.
137. Townsend, G.; LaPallo, B.K.; Boulay, C.B.; Krusienski, D.J.; Frye, G.E.; Hauser, C.K.;
Schwartz, N.E.; Vaughan, T.M.; Wolpaw, J.R.; Sellers, E.W. A novel P300-based
brain-computer interface stimulus presentation paradigm: Moving beyond rows and columns.
Clin. Neurophysiol. 2010, 121, 1109–1120.
138. Brunner, P.; Joshi, S.; Briskin, S.; Wolpaw, J.R.; Bischof, H.; Schalk, G. Does the ‘P300’ speller
depend on eye gaze? J. Neural Eng. 2010, 7, 056013.
139. Pfurtscheller, G.; Lopes da Silva, F.H. Event-related EEG/MEG synchronization and
desynchronization: Basic principles. Clin. Neurophysiol. 1999, 110, 1842–1857.
140. Jeannerod, M. Mental imagery in the motor context. Neuropsychologia 1995, 33, 1419–1432.
141. Pfurtscheller, G.; Neuper, C.; Flotzinger, D.; Pregenzer, M. EEG-based discrimination between
imagination of right and left hand movement. Electroencephalogr. Clin. Neurophysiol. 1997,
103, 642–651.
142. Blankertz, B.; Sannelli, C.; Halder, S.; Hammer, E.M.; Kübler, A.; Müller, K.R.; Curio, G.;
Dickhaus, T. Neurophysiological predictor of SMR-based BCI performance. Neuroimage 2010,
51, 1303–1309.
143. Neuper, C.; Scherer, R.; Reiner, M.; Pfurtscheller, G. Imagery of motor actions: Differential
effects of kinesthetic and visual-motor mode of imagery in single-trial EEG. Cogn. Brain. Res.
2005, 25, 668–677.
144. Nijboer, F.; Furdea, A.; Gunst, I.; Mellinger, J.; McFarland, D.J.; Birbaumer, N.; Kübler, A. An
auditory brain-computer interface (BCI). J. Neurosci. Methods 2008, 167, 43–50.
145. Hwang, H.; Kwon, K.; Im, C. Neurofeedback-based motor imagery training for brain-computer
interface (BCI). J. Neurosci. Methods 2009, 179, 150–156.
146. Wolpaw, J.R.; McFarland, D.J.; Vaughan, T.M. Brain-computer interface research at the
Wadsworth Center. IEEE Trans. Rehabil. Eng. 2000, 8, 222–226.
147. Blankertz, B.; Losch, F.; Krauledat, M.; Dornhege, G.; Curio, G.; Muller, K.R. The berlin
brain-computer interface: Accurate performance from first-session in BCI-Naïve subjects. IEEE
Trans. Biomed. Eng. 2008, 55, 2452–2462.
148. Pfurtscheller, G.; Neuper, C.; Muller, G.R.; Obermaier, B.; Krausz, G.; Schlogl, A.; Scherer, R.;
Graimann, B.; Keinrath, C.; Skliris, D.; et al. Graz-BCI: State of the art and clinical applications.
IEEE Trans. Neural Sys. Rehabil. Eng. 2003, 11, 1–4.
Sensors 2012, 12 1268
149. Bai, O.; Rathi, V.; Lin, P.; Huang, D.; Battapady, H.; Fei, D.; Schneider, L.; Houdayer, E.;
Chen, X.; Hallett, M. Prediction of human voluntary movement before it occurs. Clin.
Neurophysiol. 2011, 122, 364–372.
150. Kleber, B.; Birbaumer, N. Direct brain communication: neuroelectric and metabolic approaches
at Tübingen. Cogn. Process. 2005, 6, 65–74.
151. Krusienski, D.J.; Schalk, G.; McFarland, D.J.; Wolpaw, J.R. A mu-rhythm matched filter for
continuous control of a brain-computer interface. IEEE Trans. Biomed. Eng. 2007, 54, 273–280.
152. Tsui, C.; Gan, J.; Asynchronous BCI Control of a Robot Simulator with Supervised Online
Training. In Intelligent Data Engineering and Automated Learning—IDEAL 2007; Yin, H.,
Tino, P., Corchado, E., Byrne, W., Yao, X., Eds.; Springer: Berlin, Germany, 2007; Volume 4881,
pp. 125–134.
153. Ghafar, R.; Hussain, A.; Samad, S.A.; Tahir, N.M. Comparison of FFT and AR Techniques for
Scalp EEG Analysis. In Proceedings of the 4th Kuala Lumpur International Conference on
Biomedical Engineering 2008, Kuala Lumpur, Malaysia, June 2008; Abu Osman, N.A., Ibrahim,
F., Wan Abas, W.A.B., Abdul Rahman, H.S., Ting, H.-N., Magjarevic, R., Eds.; Springer Berlin
Heidelberg: Berlin, Germany, 2008; Volume 21, pp. 158–161.
154. von Bünau, P.; Meinecke, F.C.; Király, F.C.; Müller, K.R. Finding stationary subspaces in
multivariate time series. Phys. Rev. Lett. 2009, 103, 214101.
155. Lin, C.J.; Hsieh, M.H. Classification of mental task from EEG data using neural networks based
on particle swarm optimization. Neurocomputing 2009, 72, 1121–1130.
156. McFarland, D.J.; Anderson, C.W.; Muller, K.R.; Schlogl, A.; Krusienski, D.J. BCI meeting
2005-workshop on BCI signal processing: Feature extraction and translation. IEEE Trans.
Neural Syst. Rehabil. Eng. 2006, 14, 135–138.
157. Lins, O.G.; Picton, T.W.; Berg, P.; Scherg, M. Ocular artifacts in recording EEGs and
event-related potentials II: Source dipoles and source components. Brain Topogr. 1993, 6,
65–78.
158. Boye, A.T.; Kristiansen, U.Q.; Billinger, M.; do Nascimento, O.F.; Farina, D. Identification of
movement-related cortical potentials with optimized spatial filtering and principal component
analysis. Biomed. Signal Process. Control 2008, 3, 300–304.
159. Te-Won, L.; Lewicki, M.S.; Girolami, M.; Sejnowski, T.J. Blind source separation of more
sources than mixtures using overcomplete representations. IEEE Signal Process. Lett. 1999, 6,
87–90.
160. Kun, L.; Sankar, R.; Arbel, Y.; Donchin, E. Single trial independent component analysis for
P300 BCI system. In Proceedings of the 31th Annual International Conference of the IEEE
Engineering in Medicine and Biology Society (EMBCS’09), Minneapolis, MN, USA, September
2009; pp. 4035–4038.
161. Castellanos, N.P.; Makarov, V.A. Recovering EEG brain signals: Artifact suppression with
wavelet enhanced independent component analysis. J. Neurosci. Methods 2006, 158, 300–312.
162. Bell, A.; Sejnowski, T. An information-maximization approach to blind separation and blind
deconvolution. Neural Comput. 1995, 7, 1129–1159.
Sensors 2012, 12 1269
163. Lee, T.; Girolami, M.; Sejnowski, T.J. Independent component analysis using an extended
infomax algorithm for mixed subgaussian and supergaussian sources. Neural Comput. 1999, 11,
417–441.
164. Gao, J.; Yang, Y.; Lin, P.; Wang, P.; Zheng, C. Automatic removal of eye-movement and blink
artifacts from EEG signals. Brain Topogr. 2010, 23, 105–114.
165. Flexer, A.; Bauer, H.; Pripfl, J.; Dorffner, G. Using ICA for removal of ocular artifacts in EEG
recorded from blind subjects. Neural Netw. 2005, 18, 998–1005.
166. Jung, T.P.; Makeig, S.; Westerfield, M.; Townsend, J.; Courchesne, E.; Sejnowski, T.J. Removal
of eye activity artifacts from visual event-related potentials in normal and clinical subjects. Clin.
Neurophysiol. 2000, 111, 1745–1758.
167. Wallstrom, G.L.; Kass, R.E.; Miller, A.; Cohn, J.F.; Fox, N.A. Automatic correction of ocular
artifacts in the EEG: a comparison of regression-based and component-based methods. Int. J.
Psychophysiol. 2004, 53, 105–119.
168. Chiappa, S.; Barber, D. EEG classification using generative independent component analysis.
Neurocomputing 2006, 69, 769–777.
169. Kay, S.M.; Marple, S.L., Jr. Spectrum analysis—A modern perspective. Proc. IEEE 1981, 69,
1380-1419.
170. Krusienski, D.J.; McFarland, D.J.; Wolpaw, J.R. An Evaluation of Autoregressive Spectral
Estimation Model Order for Brain-Computer Interface Applications. In Proceedings of the 28th
Annual International Conference of the IEEE Engineering in Medicine and Biology Society
(EMBS’06), New York, NY, USA, September 2006; pp. 1323–1326.
171. Florian, G.; Pfurtscheller, G. Dynamic spectral analysis of event-related EEG data.
Electroencephalogr. Clin. Neurophysiol. 1995, 95, 393–396.
172. Jiang, W.; Guizhi, X.; Lei, W.; Huiyuan, Z. Feature Extraction of Brain-Computer Interface
Based on Improved Multivariate Adaptive Autoregressive Models. In Proceedings of the 3rd
International Conference on Biomedical Engineering and Informatics (BMEI’10), Yantai, China,
October 2010; pp. 895–898.
173. Brunner, C.; Allison, B.Z.; Krusienski, D.J.; Kaiser, V.; Müller-Putz, G.R.; Pfurtscheller, G.;
Neuper, C. Improved signal processing approaches in an offline simulation of a hybrid
brain-computer interface. J. Neurosci. Methods 2010, 188, 165–173.
174. Ghanbari, A.A.; Kousarrizi, M.R.N.; Teshnehlab, M.; Aliyari, M. Wavelet and Hilbert
transform-based Brain Computer Interface. In Proceedings of the International Conference on
Advances in Computational Tools for Engineering Applications (ACTEA’09), Beirut, Lebanon,
July 2009; pp. 438–442.
175. Samar, V.J.; Bopardikar, A.; Rao, R.; Swartz, K. Wavelet analysis of neuroelectric waveforms: A
conceptual tutorial. Brain Lang 1999, 66, 7–60.
176. Demiralp, T.; Yordanova, J.; Kolev, V.; Ademoglu, A.; Devrim, M.; Samar, V.J. Time-frequency
analysis of single-sweep event-related potentials by means of fast wavelet transform. Brain Lang
1999, 66, 129–145.
177. Hubbard, B.B. The World According to Wavelets: The Story of a Mathematical Technique in the
Making; A.K. Peters: Natick, MA, USA, 1998.
Sensors 2012, 12 1270
178. Farina, D.; do Nascimento, O.F.; Lucas, M.F.; Doncarli, C. Optimization of wavelets for
classification of movement-related cortical potentials generated by variation of force-related
parameters. J. Neurosci. Methods 2007, 162, 357–363.
179. Bostanov, V. BCI competition 2003-data sets Ib and IIb: Feature extraction from event-related
brain potentials with the continuous wavelet transform and the t-value scalogram. IEEE Trans.
Biomed. Eng. 2004, 51, 1057–1061.
180. Senkowski, D.; Herrmann, C.S. Effects of task difficulty on evoked gamma activity and ERPs in
a visual discrimination task. Clin. Neurophysiol. 2002, 113, 1742–1753.
181. Mason, S.G.; Birch, G.E. A brain-controlled switch for asynchronous control applications. IEEE
Trans. Biomed. Eng. 2000, 47, 1297–1307.
182. Hinterberger, T.; Kübler, A.; Kaiser, J.; Neumann, N.; Birbaumer, N. A brain-computer interface
(BCI) for the locked-in: Comparison of different EEG classifications for the thought translation
device. Clin. Neurophysiol. 2003, 114, 416–425.
183. Ramoser, H.; Muller-Gerking, J.; Pfurtscheller, G. Optimal spatial filtering of single trial EEG
during imagined hand movement. IEEE Trans. Rehabil. Eng. 2000, 8, 441–446.
184. Grosse-Wentrup, M.; Buss, M. Multiclass common spatial patterns and information theoretic
feature extraction. IEEE Trans. Biomed. Eng. 2008, 55, 1991–2000.
185. Mousavi, E.A.; Maller, J.J.; Fitzgerald, P.B.; Lithgow, B.J. Wavelet common spatial pattern in
asynchronous offline brain computer interfaces. Biomed. Signal Process. Control 2011, 6, 121–128.
186. Lemm, S.; Blankertz, B.; Curio, G.; Muller, K.R. Spatio-spectral filters for improving the
classification of single trial EEG. IEEE Trans. Biomed. Eng. 2005, 52, 1541–1548.
187. Dornhege, G.; Blankertz, B.; Krauledat, M.; Losch, F.; Curio, G.; Muller, K.R. Combined
optimization of spatial and temporal filters for improving brain-computer interfacing. IEEE
Trans. Biomed. Eng. 2006, 53, 2274–2281.
188. Dal Seno, B.; Matteucci, M.; Mainardi, L. A Genetic Algorithm for Automatic Feature
Extraction in P300 Detection. In Proceedings of the IEEE International Joint Conference on
Neural Networks (IJCNN’08), Hong Kong, China, June 2008; pp. 3145–3152.
189. Fatourechi, M.; Birch, G.E.; Ward, R.K. Application of a hybrid wavelet feature selection
method in the design of a self-paced brain interface system. J. NeuroEng. Rehabil. 2007, 4, 1–13.
190. Ciaccio, E.; Dunn, S.; Akay, M. Biosignal pattern recognition and interpretation systems. 2.
Methods for feature extraction and selection. IEEE Eng. Med. Biol. Mag. 1993, 12, 106–113.
191. Dias, N.; Kamrunnahar, M.; Mendes, P.; Schiff, S.; Correia, J. Feature selection on movement
imagery discrimination and attention detection. Med. Biol. Eng. Comput. 2010, 48, 331–341.
192. Lakany, H.; Conway, B.A. Understanding intention of movement from electroencephalograms.
Expert. Syst. 2007, 24, 295–304.
193. Pudil, P.; Novovičová, J.; Kittler, J. Floating search methods in feature selection. Pattern
Recogn. Lett. 1994, 15, 1119–1125.
194. Townsend, G.; Graimann, B.; Pfurtscheller, G. A comparison of common spatial patterns with
complex band power features in a four-class BCI experiment. IEEE Trans. Biomed. Eng. 2006,
53, 642–651.
Sensors 2012, 12 1271
195. Hasan, B.A.S.; Gan, J.Q. Unsupervised Adaptive GMM for BCI. In Proceedings of the 4th
International IEEE/EMBS Conference on Neural Engineering (NER’09), Antalya, Turkey, May
2009; pp. 295–298.
196. Neuper, C.; Scherer, R.; Wriessnegger, S.; Pfurtscheller, G. Motor imagery and action
observation: Modulation of sensorimotor brain rhythms during mental control of a brain-computer
interface. Clin. Neurophysiol. 2009, 120, 239–247.
197. Fatourechi, M.; Bashashati, A.; Ward, R.K.; Birch, G.E. EMG and EOG artifacts in brain
computer interface systems: A survey. Clin. Neurophysiol. 2007, 118, 480–494.
198. Croft, R.J.; Barry, R.J. Removal of ocular artifact from the EEG: A review. Neurophysiol. Clin.
2000, 30, 5–19.
199. Vigário, R.N. Extraction of ocular artefacts from EEG using independent component analysis.
Electroencephalogr. Clin. Neurophysiol. 1997, 103, 395–404.
200. del R Millan, J.; Mourino, J.; Franze, M.; Cincotti, F.; Varsta, M.; Heikkonen, J.; Babiloni, F.
A local neural classifier for the recognition of EEG patterns associated to mental tasks. IEEE
Trans. Neural Netw. 2002, 13, 678–686.
201. Changmok, C.; Micera, S.; Carpaneto, J.; Jung, K. Development and quantitative performance
evaluation of a noninvasive EMG computer interface. IEEE Trans. Biomed. Eng. 2009, 56, 188–191.
202. Nojd, N.; Hannula, M.; Hyttinen, J. Electrode Position Optimization for Facial EMG
Measurements for Human-Computer Interface. In Proceedings of the 2nd International
Conference on Pervasive Computing Technologies for Healthcare (PCTH’11), Dublin, Ireland,
February 2008; pp. 319–322.
203. Inhyuk, M.; Myoungjoon, L.; Museong, M. A Novel EMG-Based Human-Computer Interface
for Persons with Disability. In Proceedings of the Proceedings of the IEEE International
Conference on Mechatronics (ICM’04), Tunis, Tunisia, June 2004; pp. 519–524.
204. Gomez-Gil, J.; San-Jose-Gonzalez, I.; Nicolas-Alonso, L.F.; Alonso-Garcia, S. Steering a tractor
by means of an EMG-based human-machine interface. Sensors 2011, 11, 7110–7126.
205. Nan, B.; Okamoto, M.; Tsuji, T. A hybrid motion classification approach for EMG-based
human-robot interfaces using bayesian and neural networks. IEEE Trans. Robot. 2009, 25,
502–511.
206. Zhang, X.; Chen, X.; Lantz, V.; Yang, J.-H.; Wang, K.-Q. Exploration on the Feasibility of
Building Muscle-Computer Interfaces Using Neck and Shoulder Motions. In Proceedings of the
Annual International Conference of the IEEE Engineering in Medicine and Biology Society
(EMBC’09), Minneapolis, MN, USA, 3–6 September 2009; pp. 7018–7021.
207. De la Rosa, R.; Alonso, A.; Carrera, A.; Durán, R.; Fernández, P. Man-machine interface system for
neuromuscular training and evaluation based on EMG and MMG signals. 2010, 10, 11100–11125.
208. Eckhouse, R.H.; Maulucci, R.A. A multimedia system for augmented sensory assessment and
treatment of motor disabilities. Telemat. Inf. 1997, 14, 67–82.
209. Lotte, F.; Congedo, M.; Lécuyer, A.; Lamarche, F.;Arnaldi, B. A review of classification
algorithms for EEG-based brain-computer interfaces. J. Neural Eng. 2007, 4, R1.
210. McFarland, D.J.; Wolpaw, J.R. Sensorimotor rhythm-based brain-computer interface (BCI):
Feature selection by regression improves performance. IEEE Trans. Neural Syst. Rehabil. Eng.
2005, 13, 372–379.
Sensors 2012, 12 1272
211. Blankertz, B.; Muller, K.R.; Krusienski, D.J.; Schalk, G.; Wolpaw, J.R.; Schlogl, A.;
Pfurtscheller, G.; Millan, J.R.; Schroder, M.; Birbaumer, N. The BCI competition III: Validating
alternative approaches to actual BCI problems. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14,
153–159.
212. McFarland, D.J.; Krusienski, D.J.; Wolpaw, J.R.; Brain-Computer Interface Signal Processing at
the Wadsworth Center: mu and Sensorimotor Beta Rhythms. In Progress in Brain Research;
Christa, N., Wolfgang, K., Eds.; Elsevier: New York, NY, USA, 2006; Volume 159, pp. 411–419.
213. Daly, J.J.; Wolpaw, J.R. Brain-computer interfaces in neurological rehabilitation. Lancet Neurol.
2008, 7, 1032–1043.
214. Shenoy, P.; Krauledat, M.; Blankertz, B.; Rao, R.P.N.; Müller, K.-R. Towards adaptive
classification for BCI. J. Neural Eng. 2006, 3, R13.
215. Millan, J.R.; Mourino, J. Asynchronous BCI and local neural classifiers: an overview of the
adaptive brain interface project. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 159–161.
216. Galán, F.; Nuttin, M.; Lew, E.; Ferrez, P.W.; Vanacker, G.; Philips, J.; Millán, J.D.R. A
brain-actuated wheelchair: Asynchronous and non-invasive brain-computer interfaces for
continuous control of robots. Clin. Neurophysiol. 2008, 119, 2159–2169.
217. Li, Y.; Guan, C.; Li, H.; Chin, Z. A self-training semi-supervised SVM algorithm and its
application in an EEG-based brain computer interface speller system. Pattern Recogn. Lett. 2008,
29, 1285–1294.
218. Vidaurre, C.; Sannelli, C.; Müller, K.-R.; Blankertz, B. Machine-learning-based coadaptive
calibration for brain-computer interfaces. Neural Comput. 2010, 23, 791–816.
219. Shijian, L.; Cuntai, G.; Haihong, Z. Unsupervised brain computer interface based on intersubject
information and online adaptation. IEEE Trans. Neural Syst. Rehabil. Eng. 2009, 17, 135–145.
220. Yan, L.; Kambara, H.; Koike, Y.; Sugiyama, M. Application of covariate shift adaptation
techniques in brain-computer interfaces. IEEE Trans. Biomed. Eng. 2010, 57, 1318–1324.
221. Ferrez, P.W.; del R. Millan, J. Error-related EEG potentials generated during simulated
brain-computer interaction. IEEE Trans. Biomed. Eng. 2008, 55, 923–929.
222. Millan, J.R. On the Need for On-Line Learning in Brain-Computer Interfaces. In Proceedings of
the IEEE International Joint Conference on Neural Networks, (IJCNN’04), Budapest, Hungary,
July 2004; pp. 2877-2882.
223. Jain, A.K.; Duin, R.P.W.; Jianchang, M. Statistical pattern recognition: A review. IEEE Trans.
Pattern Anal. 2000, 22, 4–37.
224. Lemm, S.; Blankertz, B.; Dickhaus, T.; Müller, K.-R. Introduction to machine learning for brain
imaging. NeuroImage 2011, 56, 387–399.
225. Denoeux, T. A k-nearest neighbor classification rule based on Dempster-Shafer theory. IEEE
Trans. Syst. Man. Cyb. 1995, 25, 804–813.
226. Dudani, S.A. The distance-weighted k-nearest-neighbor rule. IEEE Trans. Syst. Man. Cyb. 1976,
SMC-6, 325–327.
227. Borisoff, J.F.; Mason, S.G.; Bashashati, A.; Birch, G.E. Brain-computer interface design for
asynchronous control applications: Improvements to the LF-ASD asynchronous brain switch.
IEEE Trans. Biomed. Eng. 2004, 51, 985–992.
Sensors 2012, 12 1273
228. Schlögl, A.; Lee, F.; Bischof, H.; Pfurtscheller, G. Characterization of four-class motor imagery
EEG data for the BCI-competition 2005. J. Neural Eng. 2005, 2, L14.
229. Kayikcioglu, T.; Aydemir, O. A polynomial fitting and k-NN based approach for improving
classification of motor imagery BCI data. Pattern Recogn. Lett. 2010, 31, 1207–1215.
230. Garrett, D.; Peterson, D.A.; Anderson, C.W.; Thaut, M.H. Comparison of linear, nonlinear, and
feature selection methods for EEG signal classification. IEEE Trans. Neural Syst. Rehabil. Eng.
2003, 11, 141–144.
231. Scherer, R.; Muller, G.R.; Neuper, C.; Graimann, B.; Pfurtscheller, G. An asynchronously
controlled EEG-based virtual keyboard: Improvement of the spelling rate. IEEE Trans. Biomed.
Eng. 2004, 51, 979–984.
232. Muller, K.R.; Anderson, C.W.; Birch, G.E. Linear and nonlinear methods for brain-computer
interfaces. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 165–169.
233. Blankertz, B.; Lemm, S.; Treder, M.; Haufe, S.; Müller, K.R. Single-trial analysis and
classification of ERP components—A tutorial. Neuroimage 2011, 56, 814–825.
234. Vidaurre, C.; Krämer, N.; Blankertz, B.; Schlögl, A. Time domain parameters as a feature for
EEG-based brain-computer interfaces. Neural Netw. 2009, 22, 1313–-1319.
235. Hoffmann, U.; Vesin, J.M.; Ebrahimi, T.; Diserens, K. An efficient P300-based brain-computer
interface for disabled subjects. J. Neurosci. Methods 2008, 167, 115–125.
236. Burges, C.J.C. A tutorial on support vector machines for pattern recognition. Data Min. Knowl.
Discov. 1998, 2, 121–167.
237. Xiang, L.; Dezhong, Y.; Wu, D.; Chaoyi, L. Combining spatial filters for the classification of
single-trial EEG in a finger movement task. IEEE Trans. Biomed. Eng. 2007, 54, 821–831.
238. Cover, T.M. Geometrical and statistical properties of systems of linear inequalities with
applications in pattern recognition. IEEE Trans. Electron. Comput. 1965, EC-14, 326–334.
239. Garcia, G.N.; Ebrahimi, T.; Vesin, J.M. Support Vector EEG Classification in the Fourier and
Time-Frequency Correlation Domains. In Proceedings of the First International IEEE EMBS
Conference on Neural Engineering (NER’03), Capri Island, Italy, March 2003; pp. 591–594.
240. Blankertz, B.; Curio, G.; Müller, K. Classifying Single Trial EEG: Towards Brain Computer
Interfacing. In Proceedings of the Advances in Neural Information Processing Systems
(ANPS’02), Vancouver, BC, Canada, December 2002; pp. 157–164.
241. Haihong, Z.; Cuntai, G.; Chuanchu, W. Asynchronous P300-based brain--computer interfaces: A
computational approach with statistical models. IEEE Trans. Biomed. Eng. 2008, 55, 1754–1763.
242. Kaper, M.; Meinicke, P.; Grossekathoefer, U.; Lingner, T.; Ritter, H. BCI competition 2003-data
set IIb: Support vector machines for the P300 speller paradigm. IEEE Trans. Biomed. Eng. 2004,
51, 1073–1076.
243. Meinicke, P.; Kaper, M.; Hoppe, F.; Heumann, M.; Ritter, H. Improving Transfer Rates in Brain
Computer Interfacing: A Case Study. In Proceedings of the Advances in Neural Inf. Proc.
Systems (NIPS’02), Vancouver, BC, Canada, December 2002; pp. 1107–1114.
244. Thulasidas, M.; Cuntai, G.; Jiankang, W. Robust classification of EEG signal for brain-computer
interface. IEEE Trans. Neural Sys. Rehabil. Eng. 2006, 14, 24–29.
Sensors 2012, 12 1274
245. Ruiting, Y.; Gray, D.A.; Ng, B.W.; Mingyi, H. Comparative Analysis of Signal Processing in
Brain Computer Interface. In Proceedings of the 4th IEEE Conference on Industrial Electronics
and Applications (ICIEA’09), Xi'an, China, May 2009; pp. 580-585.
246. Liu, G.; Huang, G.; Meng, J.; Zhang, D.; Zhu, X. Improved GMM with parameter initialization
for unsupervised adaptation of Brain-Computer interface. Int. J. Numer. Meth. BioMed. Eng.
2010, 26, 681–691.
247. Lemm, S.; Schafer, C.; Curio, G. BCI competition 2003-data set III: Probabilistic modeling of
sensorimotor mu rhythms for classification of imaginary hand movements. IEEE Trans. Biomed.
Eng. 2004, 51, 1077–1080.
248. Pires, G.; Castelo-Branco, M.; Nunes, U. Visual P300-Based BCI to Steer a Wheelchair: A
Bayesian Approach. In Proceedings of the 30th Annual International Conference of the IEEE
Engineering in Medicine and Biology Society (EMBS’08), Vancouver, Canada, August 2008;
pp. 658–661.
249. Palaniappan, R. Brain Computer Interface Design Using Band Powers Extracted During Mental
Tasks. In Proceedings of the 2nd International IEEE EMBS Conference on Neural Engineering
(NER’05), Arlington, VA, USA, March 2005; pp. 321–324.
250. Nakayaman, K.; Inagaki, K. A Brain Computer Interface Based on Neural Network with
Efficient Pre-Processing. In Proceedings of the International Symposium on Intelligent Signal
Processing and Communications (ISPACS’06), Yonago, Japan, December 2006; pp. 673–676.
251. Haselsteiner, E.; Pfurtscheller, G. Using time-dependent neural networks for EEG classification.
IEEE Trans. Rehabil. Eng. 2000, 8, 457–463.
252. Coyle, D.; McGinnity, T.M.; Prasad, G. Improving the separability of multiple EEG features for
a BCI by neural-time-series-prediction-preprocessing. Biomed. Signal Process. Control 2010, 5,
196–204.
253. Hazrati, M.K.; Erfanian, A. An online EEG-based brain-computer interface for controlling hand
grasp using an adaptive probabilistic neural network. Med. Eng. Phys. 2010, 32, 730–739.
254. Ting, W.; Guo-zheng, Y.; Bang-hua, Y.; Hong, S. EEG feature extraction based on wavelet
packet decomposition for brain computer interface. Measurement 2008, 41, 618–625.
255. Palaniappan, R.; Paramesran, R.; Nishida, S.; Saiwaki, N. A new brain-computer interface design
using fuzzy ARTMAP. IEEE Trans. Neural Sys. Rehabil. Eng. 2002, 10, 140–148.
256. Felzer, T.; Freisieben, B. Analyzing EEG signals using the probability estimating guarded neural
classifier. IEEE Trans. Neural Syst. Rehabil. Eng. 2003, 11, 361–371.
257. Kübler, A.; Birbaumer, N. Brain-computer interfaces and communication in paralysis: Extinction
of goal directed thinking in completely paralysed patients? Clin. Neurophysiol. 2008, 119,
2658–2666.
258. Betke, M.; Gips, J.; Fleming, P. The Camera Mouse: Visual tracking of body features to provide
computer access for people with severe disabilities. IEEE Trans. Neural Sys. Rehabil. Eng. 2002,
10, 1–10.
259. Miller, L.A.; Stubblefield, K.A.; Lipschutz, R.D.; Lock, B.A.; Kuiken, T.A. Improved
myoelectric prosthesis control using targeted reinnervation surgery: A case series. IEEE Trans.
Neural Sys. Rehabil. Eng. 2008, 16, 46–50.
Sensors 2012, 12 1275
260. Treder, M.S.; Blankertz, B. (C) overt attention and visual speller design in an
ERP-based brain-computer interface. Behav. Brain Funct. 2010, 6, 28.
261. Pfurtscheller, G.; Allison, B.Z.; Bauernfeind, G.n.; Brunner, C.; Solis Escalante, T.; Scherer, R.;
Zander, T.O.; Mueller-Putz, G.; Neuper, C.; Birbaumer, N. The hybrid BCI. Front. Neurosci.
2010, doi: 10.1186/1744-9081-6-28.
262. Liu, Y.; Zhou, Z.; Hu, D. Gaze independent brain-computer speller with covert visual search
tasks. Clin. Neurophysiol. 2011, 122, 1127–1136.
263. Treder, M.S.; Schmidt, N.M.; Blankertz, B. Gaze-independent brain-computer interfaces based
on covert attention and feature attention. J. Neural Eng. 2011, 8, 066003.
264. Moore, M.M. Real-world applications for brain-computer interface technology. IEEE Trans.
Neural Syst. Rehabil. Eng. 2003, 11, 162–165.
265. Birbaumer, N.; Ghanayim, N.; Hinterberger, T.; Iversen, I.; Kotchoubey, B.; Kubler, A.;
Perelmouter, J.; Taub, E.; Flor, H. A spelling device for the paralysed. Nature 1999, 398, 297–298.
266. Chambayil, B.; Singla, R.; Jha, R. Virtual Keyboard BCI Using Eye Blinks in EEG. In
Proceedings of the 6th IEEE International Conference on Wireless and Mobile Computing,
Networking and Communications (WiMob’10), Niagara Falls, ON, Canada, October 2010;
pp. 466–470.
267. Obermaier, B.; Muller, G.R.; Pfurtscheller, G. “Virtual keyboard” controlled by spontaneous
EEG activity. IEEE Trans. Neural Sys. Rehabil. Eng. 2003, 11, 422–426.
268. Silvoni, S.; Volpato, C.; Cavinato, M.; Marchetti, M.; Priftis, K.; Merico, A.; Tonin, P.;
Koutsikos, K.; Beverina, F.; Piccione, F. P300-based brain-computer interface communication:
evaluation and follow-up in amyotrophic lateral sclerosis. Front. Neurosci. 2009, 3, 60.
269. Intendix. Available online: http://www.intendix.com/ (accessed on 12 July 2011).
270. Fazel-Rezai, R.; Abhari, K. A region-based P300 speller for brain-computer interface. Can. J.
Elect. Comput. E. 2009, 34, 81–85.
271. Ahi, S.T.; Kambara, H.; Koike, Y. A dictionary-driven P300 speller with a modified interface.
IEEE Trans. Neural Syst. Rehabil. Eng. 2011, 19, 6–14.
272. Halder, S.; Rea, M.; Andreoni, R.; Nijboer, F.; Hammer, E.M.; Kleih, S.C.; Birbaumer, N.;
Kübler, A. An auditory oddball brain-computer interface for binary choices. Clin. Neurophysiol.
2010, 121, 516–523.
273. Klobassa, D.S.; Vaughan, T.M.; Brunner, P.; Schwartz, N.E.; Wolpaw, J.R.; Neuper, C.;
Sellers, E.W. Toward a high-throughput auditory P300-based brain-computer interface. Clin.
Neurophysiol. 2009, 120, 1252–1261.
274. Kübler, A.; Furdea, A.; Halder, S.; Hammer, E.M.; Nijboer, F.; Kotchoubey, B. A
brain-computer interface controlled auditory event-related potential (P300) spelling system for
Locked-in patients. Ann. N.Y. Acad. Sci. 2009, 1157, 90–100.
275. Karim, A.A.; Hinterberger, T.; Richter, J.; Mellinger, J.; Neumann, N.; Flor, H.; Kübler, A.;
Birbaumer, N. Neural internet: Web surfing with brain potentials for the completely paralyzed.
Neurorehabil. Neural Repair 2006, 20, 508–515.
276. Bensch, M.; Karim, A.A.; Mellinger, J.; Hinterberger, T.; Tangermann, M.; Bogdan, M.;
Rosenstiel, W.; Birbaumer, N. Nessi: An EEG-controlled web browser for severely paralyzed
patients. Comput. Intell. Neurosci. 2007, 2007, 71863.
Sensors 2012, 12 1276
277. Braz, G.P.; Russold, M.; Davis, G.M. Functional electrical stimulation control of standing and
stepping after spinal cord injury: A review of technical characteristics. Neuromodulation:
Technol. Neural Interface 2009, 12, 180–190.
278. Lauer, R.T.; Peckham, P.H.; Kilgore, K.L. EEG-based control of a hand grasp neuroprosthesis.
NeuroReport 1999, 10, 1767–1771.
279. Pfurtscheller, G.; Müller, G.R.; Pfurtscheller, J.; Gerner, H.J.; Rupp, R. ‘Thought’-control of
functional electrical stimulation to restore hand grasp in a patient with tetraplegia. Neurosci. Lett.
2003, 351, 33–36.
280. Hu, X.L.; Tong, K.Y.; Li, R.; Chen, M.; Xue, J.J.; Ho, S.K.; Chen, P.N. Post-stroke wrist
rehabilitation assisted with an intention-driven functional electrical stimulation (FES)-robot
system. In Proceedings of the IEEE International Conference on Rehabilitation Robotics
(ICORR’11), Zurich, Switzerland, July 2011; pp. 1–6.
281. Pfurtscheller, G.; Guger, C.; Müller, G.; Krausz, G.; Neuper, C. Brain oscillations control hand
orthosis in a tetraplegic. Neurosci. Lett. 2000, 292, 211–214.
282. Müller-Putz, G.R.; Scherer, R.; Pfurtscheller, G.; Rupp, R. EEG-based neuroprosthesis control:
A step towards clinical practice. Neurosci. Lett. 2005, 382, 169–174.
283. Buch, E.; Weber, C.; Cohen, L.G.; Braun, C.; Dimyan, M.A.; Ard, T.; Mellinger, J.; Caria, A.;
Soekadar, S.; Fourkas, A.; Birbaumer, N. Think to move: A neuromagnetic brain-computer
interface (BCI) system for chronic stroke. Stroke 2008, 39, 910–917.
284. Braun, C.; Schweizer, R.; Elbert, T.; Birbaumer, N.; Taub, E. Differential activation in
somatosensory cortex for different discrimination tasks. J. Neurosci. 2000, 20, 446–450.
285. Broetz, D.; Braun, C.; Weber, C.; Soekadar, S.R.; Caria, A.; Birbaumer, N. Combination of
brain-computer interface training and goal-directed physical therapy in chronic stroke: A case
report. Neurorehabil. Neural Repair 2010, 24, 674–679.
286. Caria, A.; Weber, C.; Brötz, D.; Ramos, A.; Ticini, L.F.; Gharabaghi, A.; Braun, C.; Birbaumer, N.
Chronic stroke recovery after combined BCI training and physiotherapy: A case report.
Psychophysiology 2011, 48, 578–582.
287. Hochberg, L.R.; Serruya, M.D.; Friehs, G.M.; Mukand, J.A.; Saleh, M.; Caplan, A.H.;
Branner, A.; Chen, D.; Penn, R.D.; Donoghue, J.P. Neuronal ensemble control of prosthetic
devices by a human with tetraplegia. Nature 2006, 442, 164–171.
288. Serruya, M.D.; Hatsopoulos, N.G.; Paninski, L.; Fellows, M.R.; Donoghue, J.P. Brain-machine
interface: Instant neural control of a movement signal. Nature 2002, 416, 141–142.
289. Millan, J.R.; Renkens, F.; Mourino, J.; Gerstner, W. Noninvasive brain-actuated control of a
mobile robot by human EEG. IEEE Trans. Biomed. Eng. 2004, 51, 1026–1033.
290. Rebsamen, B.; Burdet, E.; Cuntai, G.; Chee Leong, T.; Qiang, Z.; Ang, M.; Laugier, C.
Controlling a Wheelchair Using a BCI with Low Information Transfer Rate. In Proceedings of
the IEEE 10th International Conference on Rehabilitation Robotics (ICORR’07), Noordwijk,
The Netherlands, June 2007; pp. 1003–1008.
291. Rebsamen, B.; Cuntai, G.; Haihong, Z.; Chuanchu, W.; Cheeleong, T.; Ang, M.H.; Burdet, E.
A brain controlled wheelchair to navigate in familiar environments. IEEE Trans. Neural Syst.
Rehabil. Eng. 2010, 18, 590–598.
Sensors 2012, 12 1277
292. Iturrate, I.; Antelis, J.M.; Kubler, A.; Minguez, J. A Noninvasive brain-actuated wheelchair
based on a P300 neurophysiological protocol and automated navigation. IEEE Trans. Robot.
2009, 25, 614–627.
293. Philips, J.; del R. Millan, J.; Vanacker, G.; Lew, E.; Galan, F.; Ferrez, P.W.; Van Brussel, H.;
Nuttin, M. Adaptive Shared Control of a Brain-Actuated Simulated Wheelchair. In Proceedings
of the IEEE 10th International Conference on Rehabilitation Robotics (ICORR’07), Noordwijk,
The Netherlands, June 2007; pp. 408–414.
294. Vanacker, G.; del R. Millán, J.; Lew, E. Context-based filtering for assisted brain-actuated
wheelchair driving. Comput. Intell. Neurosci. 2007, 2007, 25130.
295. Fun of gaming: Measuring the human experience of media enjoyment. Available online:
http://fuga.aalto.fi/ (accessed on 12 July 2011).
296. Roman, K.; Benjamin, B.; Gabriel, C.; Klaus-Robert, M. The Berlin Brain-Computer Interface
(BBCI)—Towards a new communication channel for online control in gaming applications.
Multimed. Tools Appl. 2007, 33, 73–90.
297. Tangermann, M.; Krauledat, M.; Grzeska, K.; Sagebaum, M.; Blankertz, B.; Vidaurre, C.;
Müller, K. Playing Pinball with non-invasive BCI. Proc. NIPS 2008, 1641–1648.
298. Middendorf, M.; McMillan, G.; Calhoun, G.; Jones, K.S. Brain-computer interfaces based on the
steady-state visual-evoked response. IEEE Trans. Rehabil. Eng. 2000, 8, 211–214.
299. Lalor, E.C.; Kelly, S.P.; Finucane, C. Steady-state VEP-based brain-computer interface control in
an immersive 3D gaming environment. J. Appl. Signal Proc. 2005, 2005, 3156–3164.
300. Finke, A.; Lenhardt, A.; Ritter, H. The MindGame: A P300-based brain-computer interface
game. Neural Netw. 2009, 22, 1329–1333.
301. Microsoft Research—Computational User Experiences: Brain-Computer Interfaces. Available
online: http://research.microsoft.com/en-us/um/redmond/groups/cue/bci/ (accessed on 12 July
2011).
302. Angelakis, E.; Stathopoulou, S.; Frymiare, J.; Green, D.; Lubar, J.; Kounios, J. EEG
neurofeedback: A brief overview and an example of peak alpha frequency training for cognitive
enhancement in the elderly. Clin. Neuropsychol. 2007, 21, 110–129.
303. Hanslmayr, S.; Sauseng, P.; Doppelmayr, M.; Schabus, M.; Klimesch, W. Increasing individual
upper alpha power by neurofeedback improves cognitive performance in human subjects. Appl.
Psychophysiol. Biofeedback 2005, 30, 1–10.
304. Rota, G.; Sitaram, R.; Veit, R.; Erb, M.; Weiskopf, N.; Dogil, G.; Birbaumer, N. Self-regulation
of regional cortical activity using real-time fMRI: The right inferior frontal gyrus and linguistic
processing. Hum. Brain Mapp. 2009, 30, 1605–1614.
305. Sitaram, R.; Lee, S.; Ruiz, S.; Rana, M.; Veit, R.; Birbaumer, N. Real-time support vector
classification and feedback of multiple emotional brain states. Neuroimage 2011, 56, 753–765.
306. deCharms, R.C.; Maeda, F.; Glover, G.H.; Ludlow, D.; Pauly, J.M.; Soneji, D.; Gabrieli, J.D.E.;
Mackey, S.C. Control over brain activation and pain learned by using real-time functional MRI.
Proc. Natl. Acad. Sci. USA 2005, 102, 18626–18631.
307. Walker, J.E.; Kozlowski, G.P. Neurofeedback treatment of epilepsy. Child Adolesc. Psychiatr.
Clin. N. Am. 2005, 14, 163–176.
Sensors 2012, 12 1278
308. Sterman, M.; Egner, T. Foundation and practice of neurofeedback for the treatment of epilepsy.
Appl. Psychophysiol. Biofeedback 2006, 31, 21–35.
309. Strehl, U.; Leins, U.; Goth, G.; Klinger, C.; Hinterberger, T.; Birbaumer, N. Self-regulation of
slow cortical potentials: A new treatment for children with attention-deficit/hyperactivity
disorder. Pediatrics 2006, 118, 1530–1540.
310. Schneider, F.; Rockstroh, B.; Heimann, H.; Lutzenberger, W.; Mattes, R.; Elbert, T.; Birbaumer, N.;
Bartels, M. Self-regulation of slow cortical potentials in psychiatric patients: Schizophrenia.
Appl. Psychophysiol. Biofeedback 1992, 17, 277–292.
311. Schneider, F.; Heimann, H.; Mattes, R.; Lutzenberger, W.; Birbaumer, N. Self-regulation of slow
cortical potentials in psychiatric patients: Depression. Appl. Psychophysiol. Biofeedback 1992,
17, 203–214.
312. Schneider, F.; Elbert, T.; Heimann, H.; Welker, A.; Stetter, F.; Mattes, R.; Birbaumer, N.; Mann, K.
Self-regulation of slow cortical potentials in psychiatric patients: Alcohol dependency. Appl.
Psychophysiol. Biofeedback 1993, 18, 23–32.
313. Renaud, P.; Joyal, C.; Stoleru, S.; Goyette, M.; Weiskopf, N.; Birbaumer, N. Real-Time
Functional Magnetic Imaging—Brain-Computer Interface and Virtual Reality: Promising Tools
for the Treatment of Pedophilia. In Progress in Brain Research; Green, A.M., Chapman, C.E.,
Kalaska, J.F., Lepore, F., Eds.; Elsevier: New York, NY, USA, 2011; Volume 192, pp. 263–272.
314. Georgopoulos, A.P.; Karageorgiou, E.; Leuthold, A.C.; Lewis, S.M.; Lynch, J.K.; Alonso, A.A.;
Aslam, Z.; Carpenter, A.F.; Georgopoulos, A.; Hemmy, L.S.; et al. Synchronous neural
interactions assessed by magnetoencephalography: A functional biomarker for brain disorders.
J. Neural Eng. 2007, doi: 10.1088/1741-2560/4/4/001.
315. Cook, I.A.; Warren, C.; Pajot, S.K.; Schairer, D.; Leuchter, A.F. Regional Brain activation with
advertising images. J. Neurosci. Psychol. Econ. 2011, 4, 147–160.
316. Ambler, T.; Ioannides, A.; Rose, S. Brands on the brain: Neuro-images of advertising. Bus.
Strateg. Rev. 2000, 11, 17–30.
317. Vecchiato, G.; De Vico Fallani, F.; Astolfi, L.; Toppi, J.; Cincotti, F.; Mattia, D.; Salinari, S.;
Babiloni, F. The issue of multiple univariate comparisons in the context of neuroelectric brain
mapping: An application in a neuromarketing experiment. J. Neurosci. Methods 2010, 191,
283–289.
318. Ambler, T.; Braeutigam, S.; Stins, J.; Rose, S.; Swithenby, S. Salience and choice: Neural
correlates of shopping decisions. Psychol. Market 2004, 21, 247–261.
319. Ariely, D.; Berns, G.S. Neuromarketing: The hope and hype of neuroimaging in business. Nat.
Rev. Neurosci. 2010, 11, 284–292.
320. Neuromarketing, Neuroscientific Consumer Testing. NeuroFocus. Available online:
http://www.neurofocus.com/ (accessed on 28 December 2011).
321. Neuroconsult. Available online: http://www.neuroconsult.at/ (accessed on 8 October 2011).
322. Neuro Insight. Available online: http://www.neuro-insight.com/ (accessed on 8 October 2011).
323. EmSense: Quantitative Biosensory Metrics. Available online: http://www.emsense.com/
(accessed on 8 October 2011).
Sensors 2012, 12 1279
324. Fisher, C.E.; Chin, L.; Klitzman, R. Defining neuromarketing: Practices and professional
challenges. Harv. Rev. Psychiatry 2010, 18, 230–237.
© 2012 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article
distributed under the terms and conditions of the Creative Commons Attribution license
(http://creativecommons.org/licenses/by/3.0/).