Performance assessment in brain-computer interface-based augmentative and alternative communication
BioMedical Engineering OnLine volume 12, Article number: 43 (2013)
A large number of incommensurable metrics are currently used to report the performance of brain-computer interfaces (BCI) used for augmentative and alterative communication (AAC). The lack of standard metrics precludes the comparison of different BCI-based AAC systems, hindering rapid growth and development of this technology. This paper presents a review of the metrics that have been used to report performance of BCIs used for AAC from January 2005 to January 2012. We distinguish between Level 1 metrics used to report performance at the output of the BCI Control Module, which translates brain signals into logical control output, and Level 2 metrics at the Selection Enhancement Module, which translates logical control to semantic control. We recommend that: (1) the commensurate metrics Mutual Information or Information Transfer Rate (ITR) be used to report Level 1 BCI performance, as these metrics represent information throughput, which is of interest in BCIs for AAC; 2) the BCI-Utility metric be used to report Level 2 BCI performance, as it is capable of handling all current methods of improving BCI performance; (3) these metrics should be supplemented by information specific to each unique BCI configuration; and (4) studies involving Selection Enhancement Modules should report performance at both Level 1 and Level 2 in the BCI system. Following these recommendations will enable efficient comparison between both BCI Control and Selection Enhancement Modules, accelerating research and development of BCI-based AAC systems.
Augmentative and alternative communication (AAC) systems are used by individuals with communication disorders to supplement or replace speech or writing. A wide variety of AAC systems exist, ranging from picture and communication boards to speech generating devices[1, 2]. At minimum, all AAC systems require the user to produce a binary signal to indicate voluntary selection of an output option. While many access technologies exist to translate residual motor functions into an output signal, those with the most severe motor disabilities do not have any voluntary muscle control and thus cannot access AAC technologies. For these individuals, brain-computer interface (BCI) technology can be used as a form of augmentative and alternative communication (AAC). The architecture of a BCI-based AAC system can be represented by many different frameworks; herein, we model BCI-based AAC systems as two interconnected Modules, each of which is comprised of a number of functional components, depicted in Figure 1. The BCI Control Module translates a BCI user’s brain state into a logical control output. Its functional components may include a stimulus presentation paradigm which causes the BCI user to elicit particular brain states (e.g. the order of flashes in P300-based BCIs[5, 6], stimulus configuration, stimulus colour, stimulus rate); electrodes and amplifiers; feature extractors; and classification algorithms (e.g. co-adaptive calibration, adaptive online classification). A comprehensive review of the variations of each of these functional components is provided in. The BCI Control Module makes discrete selections from a system-dependent number of possible options. These selections are made independent of any semantic knowledge of the AAC interface, and the resulting logical control signal is sent to the Selection Enhancement Module. A Selection Enhancement Module translates this logical control to semantic control, using techniques ranging from direct association (e.g. one output option corresponds to one specific communicative symbol), to algorithms such as error correction and word prediction, to interface configuration (e.g. the Hex-o-Spell). Selection enhancement is not unique to the BCI field – it is employed in auto-text correction such as T9 predictive text and Swype on mobile phones, automatic speech recognition and existing AAC interfaces. These two Modules work in tandem to provide a means of communication for individuals who have severe motor impairments that limit their ability to speak and to access traditional AAC devices.
Many variations of each of the components of both BCI Control and Selection Enhancement Modules exist, and can be combined together in multiple ways to produce unique BCI-based AAC system configurations. To develop an optimal BCI-based AAC technology, researchers must be able to compare each of these configurations to assess the relative benefit of each component to the overall communication capacity of the system. In other words, the quest for the best BCI requires efficient evaluation criteria for the performance of each component of the communication system.
As is the case with the evaluation of any AAC system, the issue of where to measure performance is paramount. There are three locations, or Levels, at which BCI-based AAC performance can be measured, as depicted in Figure 1. Level 1 performance is measured directly at the output of the BCI Control Module. Here, a logical control output without semantic meaning is generated. This output is a single selection of one output from a variable number of options presented on the user interface, such as one of six targets in the Hex-o-Spell BCI, or one of 36 options in a traditional P300 Speller. Here, the effective generation of a logical control output is commonly assessed by measures of speed, accuracy, or a combination thereof, such as information transfer rate. To date, measurement of BCI performance has typically occurred at this Level. However, as BCI systems begin to explore improved user interfaces (e.g. integration of word prediction in spelling applications, innovative spelling systems, adaptive user interfaces[14, 15]), Level 2 measures of communication capacity at the output of the Selection Enhancement Module have become more common[14–17]. Level 2 measures of BCI performance account for the fact that a single selection by the user may have different degrees of “power” in terms of what it can accomplish when the logical control signal has been interpreted by the Selection Enhancement Module. These two Levels mirror where performance has typically been measured in traditional AAC systems. Rate enhancement strategies such as the physical arrangement of a display, linguistic predictive capabilities, and cues such as colour and abbreviation expansion have been extensively explored by the AAC field to overcome the rate problem of the slow productions of augmentative communicators. In the AAC field, Level 1 metrics include selections, switch activations, letters, words utterances, etc. per unit of time[19–22], and Level 2 metrics include metrics of text or selection savings[20–22], such as the Rate Index (average communication rate/selection rate). Finally, while the rarity of in-home BCIs being used by the target population have delayed the need to identify a higher-level measure of BCI performance, the AAC literature indicates that it is also possible to measure performance of a communication system, and therefore of a BCI, at the level of its impact on the user (e.g.[24, 25]). This would be considered a Level 3 measurement of BCI performance, and can be assessed by determining whether the presence of a BCI leads to fuller, richer communication with a partner, or an improved quality of life.
A large number of performance metrics have been used in BCI research studies to quantify the communication capacity of a specific BCI system. While accuracy is typically reported, this metric has several major limitations, including not accounting for time and being biased by chance performance in systems with different numbers of output states. Consequently, research groups are developing and publishing their own custom performance metrics. For example, in order to report the performance of a BCI where users were given the option of correcting mistakes that they had made in typing a sentence, Townsend et al. developed the “practical bit rate” while Jin et al. used the practical bit rate, with the addition of the “written symbol rate”.
To determine the variety of metrics in use, we conducted a literature review in Web of Science, combining the keywords “brain-computer interface” and “communication”. The search was limited to English communications in peer-reviewed journals dating between January 2005 and January 2012. Articles were included if they described the performance of synchronous, “goal selection” BCIs used by human participants for communication. According to these criteria, 72 articles were retained and included in the appraisal.
Within these 72 articles, 12 different combinations of metrics were used to describe BCI-based AAC performance. These combinations and their frequency of use are listed in Table 1.
The scope and type of metrics that have been used in the BCI field over the 7 years reviewed by the authors present two major problems. First, many of the metrics are incommensurate, precluding the comparison of results between different BCI studies. This presents a serious limitation to the growth and development that is possible in the BCI field. Second, many metrics are based on digital communication theory, and thus on assumptions that do not necessarily hold for BCI-based communication. In digital communication, a large quantity of data is typically available in order to measure performance. Due to the relatively slow communication speed of BCIs and human factors such as fatigue, it is rare that large amounts of data are available to quantify BCI performance. For example, within the P300-based BCI studies included in the surveyed literature, measures of performance were derived from an average of 30 selections, while the average P300 BCI had 36 classes. Consequently, many of the metrics that are used as gold standards in digital communication cannot necessarily be applied to the BCI field, resulting in the need for new field-specific standards to be developed.
As BCI-based AAC research continues to grow in popularity, there is a pressing need for the acceptance of standardized BCI evaluation metrics that can be used to report performance in any study using a BCI for AAC. Such metrics would enable the efficient comparison of various BCI components, accelerating the development of a practical, efficient BCI that can be used by individuals with severe motor impairments for the purposes of communication. This manuscript will compare the performance metrics that have been used for Level 1 and Level 2 measurement of BCI-based AAC performance, and recommend a standard metric for each level. Level 3 metrics will not be addressed; the interested reader is referred to the literature regarding measuring assistive technology impact on the user[91–93].
Level 1 performance metrics
Types of BCI control modules
BCI Control Modules for AAC utilize pattern recognition techniques to translate the electrical signals generated from the brain states of BCI users into the selection of one discrete option from a list of available outputs. The BCI Control Module functions under the assumption that (1) specific mental operations or (2) responses to specific sensory stimuli result in reproducible frequency or event-related potential patterns. Thus, two types of BCI Control Modules can be distinguished: (1) endogenous control modules, which respond to spontaneous control signals from the user (e.g. motor imagery to generate sensorimotor rhythms (SMRs)), and (2) exogenous control modules, which respond to control signals evoked from the user by a stimulus (e.g. event-related potentials such as the P300 response, or visually-evoked potentials (VEPs)). Effective Level 1 performance metrics should enable comparison within and between both types of BCI Control Modules.
Evaluation criteria for level 1 performance metrics
We define four criteria (described in detail below) for the evaluation of common Level 1 performance metrics. An effective Level 1 metric would be able to capture the performance of a maximal number of BCI-based AAC systems. While future systems may be created that cannot be measured with existing metrics, we offer the following criteria: the metric should have the ability to capture (1) throughput (throughput), (2) the performance of a BCI with a variable number of categorical outputs (categorical outputs), and (3) unbiased performance (unbiased). Furthermore, the metric should (4) be practically communicable to researchers and clinicians from various disciplines working in the field of BCI and practically calculable from the amounts of data typically gathered in BCI experiments (practicality).
BCI Control Modules must balance a tradeoff between system speed and system accuracy. While accuracy is commonly reported, the time per decision varies widely between different BCIs. In offline analysis, this parameter is often artificially varied to investigate the ideal settings for a system, so that the time per decision may not be fixed even within a given study. Effective Level 1 metrics must therefore capture system throughput (information per time). Metrics that report throughput can allow direct comparisons of varied BCI types; such metrics also allow comparisons between different configurations of the same BCI, such as those used to optimize parameter settings.
In any BCI system, output may be discrete or continuous; equivalently, in BCI-based AAC systems output may be (1) categorical (e.g. letters from the alphabet in a P300 speller) or (2) ordinal targets (e.g. targets in a one-dimensional SMR-based BCI). User performance in selecting ordinal targets is often measured by metrics that require inter-target distances as input (e.g. mean squared error); these metrics are thus not compatible with categorical data. By contrast, metrics designed for categorical outputs can be used with ordinal outputs, though they will ignore the extra information gained from the labels. Thus, Level 1 metrics must support categorical outputs to allow comparison between varied BCI types.
The reported performance of BCI Control Modules can be biased by three factors. 1) A variable number of discrete outcomes. P300-based BCI spellers enable the user to select from many options within a single trial (e.g. 4 options, 36 options, and 72 options), whereas some mu-rhythm based BCI selection tasks only permit selection between two discrete outputs within a single trial. Chance performance of the BCI Control Module is inversely related to the number of options. 2) Experimental bias, the marginal distribution of the intended BCI outputs determined by the experiment. In other words, the potential bias that is introduced if a BCI user is instructed to select one output option more frequently than others, as is likely to be the case during communication even outside of the laboratory. 3) Classifier bias, where a BCI Control Module preferentially selects one class over others. To enable efficient comparison between different BCI Control Modules, Level 1 metrics must be unbiased by any of these factors.
An effective metric must be practicably communicable between various research groups and accessible to individuals from various disciplinary backgrounds working in the BCI field. The metric must present BCI performance in a form that is practical for journal articles and sufficiently simple to be understood by those without engineering expertise. The metric must also be practically calculable from the relatively small amounts of data that are gathered from AAC-based BCI experiments with human participants. This final point is particularly significant, as communication systems such as spellers tend to have a large number of output options (classes), leading to a small number of examples per class in a typical experiment.
Common level 1 performance metrics
In light of the four criteria defined above, we present and discuss five common Level 1 performance metrics in this section: (1) error rate or classification accuracy; (2) Cohen’s Kappa coefficient; (3) confusion matrix; (4) mutual information; and (5) information transfer rate or bit rate. The discussion is summarized in Table 2. It is possible to address the limitations of some metrics through relatively minor adjustments (e.g. in addition to accuracy, one can report time per sequence and time between characters in a P300-speller BCI to enable the derivation of ITR). However, as these metrics are often reported without the information necessary for such conversions, they will be evaluated according to the four criteria under the assumption that no further information about performance is provided. Several other Level 1 BCI performance metrics exist that are typically used to measure continuous BCI output, such as the correlation coefficient and mean square error. These metrics are often used in SMR-based BCIs, but as they cannot be used with the categorical output generated by some BCI-based AAC systems, they will not be discussed further in this paper.
Error rate or classification accuracy 
This metric determines how often the BCI makes a correct selection; in other words, the percentage of total selections that are correct. While it is the most intuitive metric of BCI performance, it does not account for time, often suggesting that BCI performance increases monotonically with time per decision. Furthermore, this metric is biased by the chance performance of AAC configurations with different numbers of discrete outcomes, and assumes the existence of a single accuracy which is uniform across all possible outputs.
Cohen’s Kappa is a measure of the agreement of two observers; for a BCI-based AAC system, it is used as a measure of agreement between the correct output and the BCI Control Module output. Like classification accuracy, this metric does not account for the time required to make a selection, and does not give a measure of throughput. While Cohen’s Kappa is designed to account for chance agreement, it can produce unexpected results if the underlying class distribution is biased.
Confusion matrix 
For BCI-based AAC systems, a confusion matrix is a matrix with correct output as rows, BCI Control Module outputs as columns, and the number of occurrences in the intersections. The diagonal therefore represents the number of correct outputs. The confusion matrix does not account for time and thus does not measure throughput. The relative sums of the rows of the matrix reveals the frequency of each intended output in the experiment, thus, confusion matrices can show patterns in the error distribution (e.g. row-column errors in the P300 Speller).
The confusion matrix is often not practically calculable. Every entry in the matrix is proportional to a probability density estimate for a particular combination of correct and actual outputs; the number of density estimates that are required thus grows as the square of the number of states. Particularly in P300 experiments where 36 or more possible outputs are typical, this amount of data is rarely available. Also, this metric is often not practically communicable. While confusion matrices are 2-D when representing the performance of one specific BCI configuration; representing the performance across a varying number of stimulus presentations requires reporting a 3-D matrix. In addition, while the 2-D matrices are easily reported for BCIs with a small number of total possible outputs, they become impractically large for BCIs with a large number of total possible outputs. The classical Farwell and Donchin 36-class P300 speller would a priori require reporting a matrix with 1296 entries. Since that particular implementation makes separate decisions for rows and columns, reporting two separate, 36-element confusion matrices may be sufficient, though this requires an assumption of independence and makes interpretation more difficult. Some modern spellers, such as Townsend’s checkerboard speller, do not share this structure and would thus require reporting the full N2 entries (5184 for the Townsend speller). Furthermore, most of the entries are small in value, and therefore difficult to measure accurately. The combination of these factors makes the confusion matrix impractical.
Mutual information is a measure of the overlap between the correct output and the output of the BCI Control Module; it is a measure, in bits, of the throughput of information from the BCI. Since its formula includes marginal and error probabilities, it is robust with respect to experimental and system bias. However, to account for these sources of bias, the calculation of mutual information requires the estimation of the joint statistical distribution of the input and output; the amount of information needed for this estimation scales as the square of the number of total possible outputs of the system, making it impractical for use in a realistic setting with a BCI with a high number of possible outputs, such as a P300 Speller.
Nykopp’s bit rate is equivalent to mutual information, though this metric does not require the explicit estimation of joint statistical distribution. While Nykopp recommended maximizing the throughput measure by taking the maximum mutual information across all marginal input probabilities, this is an artificial approach, and the metric is only valid provided this maximization is not performed.
Wolpaw’s Information transfer rate (ITR) or bit rate .
Information transfer rate (ITR), also called bit rate, measures the amount of information passing through a device per unit time. It is derived from mutual information, thus, it works with categorical outputs. It is worthwhile to note that in the derivation, Wolpaw et al. assumed that the probability of error would be uniform across all possible outputs, and that errors would be uniformly distributed among the available choices. The violation of these assumptions can produce unexpected results, as shown in. However, these assumptions dramatically limit the amount of data necessary to calculate the metric.
Level 1 performance metric recommendations
As illustrated in Table 2, none of the metrics that are currently used to report performance of a BCI Control Module satisfy all four criteria of an effective Level 1 metric. Mutual information and information transfer rate both satisfy three of the four evaluation criteria. Each has a different strength: mutual information accounts for bias, while information transfer rate is more practical to calculate in light of typical data limitations of BCI experiments. Fortunately, as ITR was derived from mutual information, they are commensurable metrics. Thus, the decision of which of these two metrics to use can be guided by data availability. Standard formulas can be used to calculate the confidence bounds of each class, since the accuracy of each class is observed as a binomial random variable. If sufficient data are available to demonstrate a significant difference between classes, mutual information is preferred. However, if sufficient data are not available, the error from using poor per-class estimates of accuracy may be worse than the error expected from assuming a single uniform accuracy. We therefore recommend the use of mutual information when significant bias is expected or deliberately introduced, and the ITR approximation in other situations, as the standard metrics to report Level 1 performance of any BCI Control Modules.
Level 2 performance metrics
Types of selection enhancement modules
Three types of Selection Enhancement Modules can be defined based on their respective mechanisms for enhancing the logical output they receive from the BCI Control Module: (1) automatic error correction; (2) rate enhancement; and (3) control state detection.
Automatic error correction mechanisms are ways by which the system can recover from errors. These mechanisms include techniques such as the using the detection of an “error potential” brainwave to automatically undo erroneous selections , or the replacement of a deleted character with a likely second candidate upon the selection of a backspace .
Rate enhancement mechanisms map the discrete logical selection received from the BCI Control Module into selections with a larger unit of semantic meaning. Such mechanisms range from populating selection options with communicative signs and symbols, e.g. Blissymbols , to enhancing a P300 speller with word prediction, which enables users to complete full words in a single selection (e.g. ).
Control state detection mechanisms monitor the attention of the user, and abstain from making a selection when the user is not paying attention to the BCI, thus preventing selections which are likely to be erroneous .
While all three mechanisms operate on different principles, they are each designed with the common purpose of enhancing the effectiveness of BCI-based communication beyond what is possible with a BCI Control Module alone.
Evaluation criteria for level 2 performance metrics
In order to enable BCI researchers and users to make informed choices of the best Selection Enhancement Modules, it is imperative to have a metric that allows comparison within and across all three Selection Enhancement mechanisms described in section “Types of Selection Enhancement Modules”. In addition, the metric should be usable with different subject instructions regarding the handling of errors specified in an experimental protocol (e.g. user required to correct errors, user required to ignore errors). Finally, the criterion of practicality as defined for Level 1 metrics also apply for Level 2 performance metrics. The practicality criterion is even more important for Level 2 metrics than for Level 1 metrics, as it is likely that a broader range of disciplines will be interested in Level 2 measures of performance. Thus, a Level 2 metric must be compatible with the three types of Selection Enhancement mechanisms - 1) automatic error correction, 2) rate enhancement and 3) control state detection. It should also be compatible with 4) experimental protocols with and without error correction by the user; and should 5) be practicality calculable and communicable.
Common level 2 performance metrics
In the current BCI literature, six Level 2 metrics are used: (1) the written symbol rate, (2) practical bit rate (3) the extended confusion matrix, (4) the system efficiency, denoted - “Effsys”, (5) output characters per minute and (6) the BCI-Utility metric. We describe and discuss these metrics with respect to the five criteria presented in section “Evaluation criteria for level 2 performance metrics”. The results of the comparison are presented in Table 3.
(1) Written symbol rate 
Written symbol rate (WSR) is primarily applicable to error correction mechanisms in Selection Enhancement Modules. The formula accounts for the cost of selecting an erroneous character – selecting a backspace, then selecting the correct character – and for the fact that each selection involved in correcting the error is subject to error itself. However, the WSR strictly underestimates system performance, especially for low accuracies, as the formula uses ITR to derive the symbol rate. ITR already includes theoretical error correction; thus WSR accounts for each error twice, making it an invalid measure.
(2) Practical bit rate 
Like the WSR, practical bit rate is primarily applicable to error correction mechanisms in Selection Enhancement Modules. To accurately represent real-world communication scenarios, the formula adds a penalty of two additional selections for every error incurred, accounting for the same likelihood of making an error during the correcting process as in the original attempt. The formula used to calculate this metric is the same as that used for the BCI-Utility metric in the case where error correction is performed through a backspace entry in the matrix, but is less flexible in that no alternative form exists for other scenarios.
(3) Extended confusion matrix 
The extended confusion matrix (ECM) is an extension of the confusion matrix (described in section “Common level 1 performance metrics”) that accounts for abstentions, or situations where the BCI system deliberately decides not to output a selection. However, it requires the collection of sufficient data to provide estimates of each probability of misclassification. Thus, like mutual information or confusion matrices among the Level 1 metrics, ECM requires more information than is available in many BCI experiments (e.g. ECMs for spellers could require thousands of entries, and at best are impractical to both report and interpret).
The problem of impracticality could be reduced by reporting aggregate data from all subjects; however, this approach introduces subtle biases into the data. As examples: the backspace option is more likely to be selected by individuals with poor performance; in time-limited trials, only participants with good performance will complete the sentence, thus characters appearing earlier in the sentence are likely to show a bias towards poor performance. These subtle factors mean that even aggregate data must be reported and interpreted with caution.
ECM also does not currently have an explicit mechanism for capturing selection enhancements such as word prediction or symbolic communication. It may be possible to derive an extension that could capture these enhancements, however such a derivation is likely to require considerable effort.
(4) Effsys 
Effsys is a measure of the efficiency of a BCI system. Effsys is based on ECM, but differs in that it (a) includes calculations for the cost of errors; and (b) is a scalar metric, and therefore practical for publication. Effsys is designed to account for the fact that different outputs may have different probabilities of correct classification; however, its derivation assumes that the probability of selecting a ‘backspace’ option is equal to the probability of selecting each of the other outputs. This inconsistency leads to erratic behavior in this metric; if the accuracy of even one potential output is less than 50%, Effsys = 0. This behavior can be corrected by a slight modification to the formula, which we present as Effsys’.
Effsys’ is a modification of Effsys that accounts for the fact that different outputs may have different probabilities of correct classification. The formula presented in Table 3 is derived for the conditions of: (1) a BCI user selecting outputs (e.g. letters) with the option of undoing erroneous selections; and (2) an erroneous selection requiring the selection of an ‘undo’ option, followed by reselecting the intended output. The second condition is not true in the case of an erroneous selection of the ‘undo’ option; consequently, Effsys’ will slightly underestimate BCI performance. This formula allows all outputs except the ‘undo’ option to have any non-zero probability; only the ‘undo’ option is required to have accuracy greater than 50%. Note that this form of the metric does not account for Selection Enhancement Modules that implement automatic error correction, though it approximates user-based error correction better than Effsys.
(6) Output characters per minute 
Output characters per minute (OCM) is calculated by dividing the final length of the error-corrected output by the time required to accomplish the task. The metric has the ability to capture the performance of all three types of Selection Enhancement Modules. However, as currently presented, the metric is not applicable to experimental protocols where errors remain in the final text, and it unduly penalizes system performance in situations where users did not notice an error immediately and continued typing before returning to correct the mistake. Furthermore, OCM is restricted to character-based communication. Although the majority of BCI research has focused on spelling applications, there are a wide variety of AAC systems that take advantage of symbolic or pictorial communication. To improve communication efficiency, BCI-based AAC systems will likely adapt these well-established conventions from the AAC field; the performance of such BCI systems is impossible to capture with OCM.
(7) BCI-Utility metric 
The BCI-Utility metric is the ratio of the expected benefit per selection and the expected time per selection. The expected benefit may be greater than one, as in the case of Selection Enhancement Modules such as word-prediction; equal to one, as in the case of direct association of a single selection to a single letter; or less than one, as in the case of the Hex-o-Spell BCI where the selection of a single hex is one of two selections necessary to generate an output letter. BCI-Utility is able to effectively measure the performance of all three types of Selection Enhancement Modules, and is applicable to experimental protocols with and without error correction. Dal Seno et al. present several forms of the metric; in Table 3, we present the most general form of the metric. If none of the presented forms are appropriate, researchers may derive a new form from its basic principles to account for the specific implementation of their BCI-based AAC system. A simple example would be adding a modification for non-uniform accuracy across all possible outputs. Consequently, the BCI-Utility metric can be extended to performance enhancements that its creators did not anticipate. Indeed, the BCI-Utility metric may also be appropriate for BCIs designed for purposes other than communication. Unfortunately, as standard formulas are not available for several common scenarios, there is some possibility that errors in derivation by research groups could lead to conflicting results. While this could be addressed by a set of standard guidelines for the use of the metric, these do not currently exist; thus, the practicality of calculating the metric is not ideal at present.
Comparison of level 2 performance metrics
To further illustrate the differences between these seven Level 2 metrics, a comparison is provided in Figure 2. Data was collected in the 3-session experiment performed in. Briefly, participants (n = 22, including 9 with amyotrophic lateral sclerosis) were asked to copy a total of 9 sentences, each 23 characters in length, using a classic P300 Speller BCI. Stimuli were presented with a 31.25 ms flash, an inter-flash interval of 125 ms, and 3.5 seconds between characters. A least-squares classifier was used to determine the logical control output of this BCI Control Module. A simple direct-association Selection Enhancement Module mapped the logical control output to a single alphanumeric character, which was displayed on the computer screen. Participants corrected errors using a backspace option in the BCI. Sentences were excluded if the participant did not complete the full sentence, correcting all errors, within 15 minutes. The 75 sentences with the lowest OCM are reported in Figure 2.
In this dataset, OCM and BCI-Utility differed only on two data points (circled). In both cases, the user noticed an error only after typing several correct letters, and had to erase those letters to correct the mistake. Thus, the BCI-Utility represents an estimate of the OCM the system would have achieved without user error. WSR severely underestimates performance, while ITR overestimates performance (particularly for low accuracies), and Effsys estimates the performance of 40 of the sentences to be zero. Both Effsys and Effsys’ vary around OCM; this behavior demonstrates that for small datasets, the error from assuming uniform accuracy across classes is smaller than the measurement error on individual accuracy estimates. The ECM for the first observation of Figure 2a is presented in Figure 2b; the size and the sparsity of the matrix for a single observation illustrates the lack of practicality of using this metric to report BCI-based AAC performance.
Level 2 performance metric recommendations
As illustrated in Table 3, none of the metrics that are currently used to report performance of a BCI Selection Enhancement Module satisfy all five criteria of an effective Level 2 metric. The BCI Utility-metric and OCM both satisfy four of the five evaluation criteria. OCM has several limitations: it cannot be calculated if errors remain in the text, thus, it is incompatible with a large number of experiments that do not require user-driven error correction; it is also unable to measure performance from symbolic communication. The BCI-Utility metric is compatible with all types of Selection Enhancement Modules and experimental protocols, but has limited practicality in that it does not have a fixed form and needs to be derived for particular problems. This limitation is easily addressable through future development of standards that delineate specific formulas and guidelines for the BCI-Utility metric under varying experimental paradigms. Thus, we recommend the use of the BCI-Utility metric as the standard to report Level 2 performance of any BCI, enabling the efficient comparison of Selection Enhancement Modules. Level 1 metrics should also be reported in any research involving Selection Enhancement, so that the effect of the Selection Enhancement Module can be clearly seen, and underlying experimental differences due to different BCI Control Modules can be identified.
The continued popularity of research in and development of BCIs has created a pressing need for the adoption of standardized BCI evaluation metrics that can be used to report performance for BCI-based AAC systems. Without such metrics, BCI studies that demonstrate the performance of various BCI Control Module or Selection Enhancement Module components remain incommensurable, preventing comparisons of BCI function between labs. This severely limits progress toward developing a practical, efficient BCI that can be used for communication by individuals with severe motor impairments. Based on criteria chosen to maximize comparability between all variations of BCI-based AAC systems, we make the following recommendations:
Using mutual information/information transfer rate (ITR) as the standard metric for reporting Level 1 BCI performance, and the BCI-Utility metric as the standard metric for reporting Level 2 BCI performance.
Supplementing these standard metrics with specific metrics typically used for a particular BCI paradigm. For example, in the P300-Speller BCI, the accuracy of the system versus the number of stimulus presentations is typically reported; in this situation, we recommend reporting accuracy versus time, with ITR overlaid on top, as presented in Figure 3. This figure is commonly used for selecting speed of operation, and would not be possible without a practical metric for measuring throughput. Note that BCI-Utility should also be reported. Such a graph is not applicable for endogenous BCIs such as those controlled by SMRs, where the BCI user is presented with constant feedback; for these systems, reporting accuracy, ITR, and BCI-Utility of the system using online settings is sufficient.
Reporting both Level 1 and Level 2 metrics in all BCI studies, but particularly Selection Enhancement Module studies. The performance of BCI systems with Selection Enhancement Modules is dependent upon the performance of the BCI Control Module as well as the performance of the Selection Enhancement Module. Reporting both metrics enables the performance of each module to be assessed independently. Similarly, when BCI systems are eventually assessed at the level of the user, it will be important to report Level 1, Level 2 and Level 3 metrics simultaneously, so that effective comparisons can be drawn between different BCI systems.
While the recommended metrics enable efficient comparison of most existing BCI systems, they may be limited in their ability to measure the performance of BCI systems that are developed under control paradigms other than those mentioned in this paper. For example, theoretical self-paced BCIs are continuously available to the user, and aware of when the user is engaging with the BCI interface or paying attention to something else (i.e. they support no-control states)[110, 111]. Neither of the recommended metrics would be adequate to measure performance of such a system. Further, throughput is not necessarily a key metric for self-paced BCIs, which incorporate potentially long periods of subject inaction. Depending on the application, other metrics may be more suitable. For example, in applying a BCI to operate a call bell in a hospital environment, a metric such as precision-recall curves or receiver-operator characteristics (ROC) may be appropriate. However, if self-paced BCIs are used to communicate frequently, this would likely be accomplished through a scanning system where options were selected via a single switch (e.g.). In this case, the BCI-Utility metric could be extended through careful measurement of the average time per selection and accuracy of selection achieved.
The recommendations in section “Recommendations” are specific to BCIs used for AAC, which are goal selection BCIs. Process control BCIs, such as those used for the purposes of mobility or environmental control (e.g. to drive a power wheelchair or to operate a call-button) may use different evaluation criteria to select efficient performance metrics, as system accuracy is often more important than throughput. In such situations, the benefit of each selection (a critical concept in BCI-Utility) may be difficult to define. Therefore, while we recommend reporting the above metrics in any BCI research that includes communication, we do not expect the metrics to capture all aspects of system performance outside the realm of BCI-based AAC systems.
The selection of standard metrics to report Level 1 and Level 2 BCI-based AAC performances is a critical first step in enabling effective comparison of various BCI systems used for communication. The adoption of these metrics as the standard in the field is necessary, but not sufficient, to achieve this goal. A set of guidelines must also be established within the BCI field that detail the appropriate ways of presenting and using each of the metrics recommended in this review. Examples of issues to be resolved in future guidelines are: ITR has sometimes been reported with the pause between characters in the P300 Speller removed, which makes comparison between studies difficult; BCI-Utility metric is only effective when comparing symbol-based versus letter-based selections if the relative benefit of symbolic-based communication is provided. This review provides a foundation for the development of such guidelines; future work in this direction is encouraged in order to develop widely-accepted standards that are used to report BCI-based AAC performance using these recommended metrics. Of particular importance is the development of guidelines for use of the BCI-Utility metric and example derivations that extend those presented in. While this is the only level 2 metric capable of capturing and comparing all of the Selection Enhancement research currently being conducted with BCIs, it currently runs the risk of leading to confusion or misinterpretation if different definitions of the “benefit” of a selection are used. Finally, as BCIs transition from laboratory-based technologies to home-based technologies, the development of standard Level 3 metrics will be necessary to facilitate the comparison and development of effective BCI-based AAC systems that can be used by individuals with severe motor impairments in a naturalistic communication setting.
Finally, it is important to recognize that in spite of our best efforts, there are experimental factors that potentially bias comparisons that cannot be corrected for by any single metric. Information about performance is always obtained under a restricted set of parameters that may favor one device over another. Standardizing the metrics used by the BCI field is advantageous to all involved, however, researchers must be vigilant against the biases inherent in each metric to ensure fair comparison of the performance of different BCI systems.
Based on the criteria proposed in this paper, we recommend that when results of BCI-based AAC studies are disseminated: (1) Mutual Information or ITR should be used to report Level 1 BCI performance, depending on the amount of data available and the presence of bias, and the BCI-Utility metric should be used to report Level 2 BCI performance; (2) these metrics should be supplemented by information specific to each unique BCI configuration (see Figure 3 as an example); and (3) studies involving Selection Enhancement Modules should report performance at both Level 1 and Level 2 in the BCI system. Following these recommendations will enable efficient comparison between both BCI Control and Selection Enhancement Modules, accelerating the development of a practical, efficient BCI that can be used by individuals with severe motor impairments for the purposes of communication.
Venkatagiri H: Clinical implications of an augmentative and alternative communication taxonomy. Augment Altern Commun 2002, 18: 45–57.
Glennen SL, DeCoste DC: Augmentative and Alternative Communication Systems. In The Handbook of Augmentative and Alternative Communication. San Diego, CA: Cengage Learning; 1997.
Tai K, Blain S, Chau T: A review of emerging access technologies for individuals with severe motor impairments. Assist Technol 2008, 20: 204–219. 10.1080/10400435.2008.10131947
Mason SG, Birch GE: A general framework for brain-computer interface design. IEEE Trans Neural Syst Rehabil Eng 2003, 11: 70–85. 10.1109/TNSRE.2003.810426
Farwell LA, Donchin E: Talking off the top of your head: toward a mental prosthesis utilizing event-related brain potentials. Electroencephalogr Clin Neurophysiol 1988, 70: 510–523. 10.1016/0013-4694(88)90149-6
Townsend G, LaPallo B, Boulay C, Krusienski D, Frye G, Hauser C, Schwartz N, Vaughan T, Wolpaw J, Sellers E: A novel P300-based brain-computer interface stimulus presentation paradigm: moving beyond rows and columns. Clin Neurophysiol 2010, 121: 1109–1120. 10.1016/j.clinph.2010.01.030
Jin J, Allison BZ, Sellers EW, Brunner C, Horki P, Wang XY, Neuper C: An adaptive P300-based control system. J Neural Eng 2011., 8: 036006
Takano K, Komatsu T, Hata N, Nakajima Y, Kansaku K: Visual stimuli for the P300 brain-computer interface: A comparison of white/gray and green/blue flicker matrices. Clin Neurophysiol 2009, 120: 1562–1566. 10.1016/j.clinph.2009.06.002
McFarland DJ, Sarnacki WA, Townsend G, Vaughan T, Wolpaw JR: The P300-based brain-computer interface (BCI): Effects of stimulus rate. Clin Neurophysiol 2011, 122: 731–737. 10.1016/j.clinph.2010.10.029
Vidaurre C, Sannelli C, Muller KR, Blankertz B: Co-adaptive calibration to improve BCI efficiency. J Neural Eng 2011., 8: 025009
Vidaurre C, Schlogl A, Cabeza R, Scherer R, Pfurthscheller G: A fully on-line adaptive BCI. IEEE Trans Biomed Eng 2006, 53: 1214–1219. 10.1109/TBME.2006.873542
Mason SG, Bashashati A, Fatourechi M, Navarro KF, Birch GE: A Comprehensive Survey of Brain Interface Technology Designs. Ann Biomed Eng 2006, 35: 137–169.
Blankertz B, Dornhege G, Krauledat M, Schoroder M, Williamson J, Murray-Smith R, Muller KR: The Berlin Brain-Computer Interface Presents the Novel Mental Typewriter Hex-O-Spell. 2006. [3rd International Brain-Computer Interface Workshop and Training Course]
Ryan DB, Frye GE, Townsend G, Berry DR, Mesa-G S, Gates NA, Sellers EW: Predictive spelling with a P300-Based brain-computer interface: increasing the rate of communication. Int J Human-Computer Inter 2011, 27: 69–84.
Furdea A, Halder S, Krusienski DJ, Bross D, Nijboer F, Birbaumer N, Kubler A: An auditory oddball (P300) spelling system for brain-computer interfaces. Psychophysiology 2009, 46: 617–625. 10.1111/j.1469-8986.2008.00783.x
Dal Seno B, Matteucci M, Mainardi LT: The utility metric: a novel method to assess the overall performance of discrete brain-computer interfaces. IEEE Trans Neural Syst Rehabil Eng 2010, 18: 20–29.
Bianchi L, Quitadamo LR, Garreffa G, Cardarilli GC, Marciani MG: Performances evaluation and optimization of brain computer interface systems in a copy spelling task. IEEE Trans Neural Syst Rehabil Eng 2007, 15: 207–216.
Beukelman DR, Mirenda P: Augmentative and Alternative Communication: Management of Severe Communication Disorders in Children and Adults. Baltimore, MD: Paul H Brookes; 1992.
Goodenough-Trepagnier C, Rosen MJ: Predicitive assessmet for communication aid prescription: Motor-determined maximum communication rate. In The Vocally Impaired: Clinical Practice and Research. Edited by: Bernstein L. Philadelphia: Grune & Stratton; 1988.
Koester HH, Levine SP: Learning and performance of able-bodied individuals using scanning systems with and without word prediction. Assist Technol 1994, 6: 42–53. 10.1080/10400435.1994.10132226
Koester HH, Levine SP: Modeling the speed of text entry with a word prediction interface. IEEE Trans Rehabil Eng 1994, 2: 177–187. Sep 10.1109/86.331567
Lesher G, Moulton B: Techniques for augmenting scanning communication. Augment Altern Commun 1998, 14: 81–81. 10.1080/07434619812331278236
Hill K, Romich B: A rate index for augmentative and alternative communication. Int J Speech Technol 2002, 5: 57–64. 10.1023/A:1013638916623
Waller A, Dennis F, Brodie J, Cairns AY: Evaluating the use of TalksBac, a predictive communication device for nonfluent adults with aphasia. Int J Lang Commun Disord 1998, 33: 45–70. 10.1080/136828298247929
Scherer MJ, Gray DB, Quatrano LA, Lieberman ML: The impact of assistive technology on the lives of people with disabilities. In Designing and Using Assistive Technology: The Human Perspective. Balitmore, MD: Paul H. Brookes Publishing Co; 1996:99–115.
Wolpaw JR: Brain–computer interfaces as new brain output pathways. J Physiol 2007, 579: 613–619. 10.1113/jphysiol.2006.125948
Friedrich EVC, Scherer R, Sonnleitner K, Neuper C: Impact of auditory distraction on user performance in a brain-computer interface driven by different mental tasks. Clin Neurophysiol 2011, 122: 2003–2009.
Hsu WY, Lin CY, Kuo WF, Liou M, Sun YN, Tsai ACH, Hsu HJ, Chen PH, Chen IR: Unsupervised fuzzy c-means clustering for motor imagery EEG recognition. Int J Innovative Comput Inform Control 2011, 7: 4965–4976.
Hsu WY: EEG-based motor imagery classification using enhanced active segment selection and adaptive classifier. Comput Biol Med 2011, 41: 633–639. 10.1016/j.compbiomed.2011.05.014
Bobrov P, Frolov A, Cantor C, Fedulova I, Bakhnyan M, Zhavoronkov A: Brain-Computer Interface Based on Generation of Visual Images. PLoS One 2011, 6: e20674. 10.1371/journal.pone.0020674
Arvaneh M, Guan CT, Ang KK, Quek C: Optimizing the channel selection and classification accuracy in EEG-Based BCI. IEEE Trans Biomed Eng 2011, 58: 1865–1873.
Ikegami S, Takano K, Saeki N, Kansaku K: Operation of a P300-based brain-computer interface by individuals with cervical spinal cord injury. Clin Neurophysiol 2011, 122: 991–996. 10.1016/j.clinph.2010.08.021
Krusienski DJ, Shih JJ: Control of a visual keyboard using an electrocorticographic brain-computer interface. Neurorehabil Neural Repair 2011, 25: 323–331. 10.1177/1545968310382425
Cecotti H, Rivet B, Congedo M, Jutten C, Bertrand O, Maby E, Mattout J: A robust sensor-selection method for P300 brain-computer interfaces. J Neural Eng 2011., 8: 016001
Li YQ, Nam CS, Shadden BB, Johnson SL: A P300-Based Brain-Computer Interface: Effects of Interface Type and Screen Size. Int J Hum-Comput Interact 2011, 27: 52–68.
Brunner P, Joshi S, Briskin S, Wolpaw JR, Bischof H, Schalk G: Does the “P300” speller depend on eye gaze? J Neural Eng 2010, 7: 056013. 10.1088/1741-2560/7/5/056013
Sellers EW, Vaughan TM, Wolpaw JR: A brain-computer interface for long-term independent home use. Amyotroph Lateral Scler 2010, 11: 449–455. 10.3109/17482961003777470
Hashimoto Y, Ushiba J, Kimura A, Liu MG, Tomita Y: Change in brain activity through virtual reality-based brain-machine communication in a chronic tetraplegic subject with muscular dystrophy. BMC Neurosci 2010, 11: 9. 10.1186/1471-2202-11-9
Zhang D, Maye A, Gao XR, Hong B, Engel AK, Gao SK: An independent brain-computer interface using covert non-spatial visual selective attention. J Neural Eng 2010., 7: 016010
Cabrera AF, Farina D, Dremstrup K: Comparison of feature selection and classification methods for a brain-computer interface driven by non-motor imagery. Med Biol Eng Comput 2010, 48: 123–132. 10.1007/s11517-009-0569-2
Guger C, Daban S, Sellers EW, Holzner C, Krausz G, Carabalona R, Gramatica F, Edlinger G: How many people are able to control a P300-based brain-computer interface (BCI)? Neurosci Lett 2009, 462: 94–98. 10.1016/j.neulet.2009.06.045
Fazel-Rezai R, Abhari K: A region-based P300 speller for brain-computer interface. Can J Elect Comput Eng 2009, 34: 81–85.
Kayagil TA, Bai O, Henriquez CS, Lin P, Furlani SJ, Vorbach S, Hallett M: A binary method for simple and accurate two-dimensional cursor control from EEG with minimal subject training. J Neuroeng Rehabil 2009., 6: 10.1186/1743-0003-6-14
Friedrich EVC, McFarland DJ, Neuper C, Vaughan TM, Brunner P, Wolpaw JR: A scanning protocol for a sensorimotor rhythm-based brain-computer interface. Biol Psychol 2009, 80: 169–175. 10.1016/j.biopsycho.2008.08.004
Sano A, Bakardjian H: Movement-related cortical evoked potentials using four-limb imagery. Int J Neurosci 2009, 119: 639–663. 10.1080/00207450802325561
Guo F, Hong B, Gao X, Gao S: A brain-computer interface using motion-onset visual evoked potential. J Neural Eng 2008, 5: 477–485. 10.1088/1741-2560/5/4/011
Blankertz B, Losch F, Krauledat M, Dornhege G, Curio G, Muller K-R: The Berlin brain–computer interface: accurate performance from first-session in BCI-NaÏve subjects. IEEE Trans Biomed Eng 2008, 55: 2452–2462.
Bai O, Lin P, Vorbach S, Floeter MK, Hattori N, Hallett M: A high performance sensorimotor beta rhythm-based brain-computer interface associated with human natural motor behavior. J Neural Eng 2008, 5: 24–35. 10.1088/1741-2560/5/1/003
Chatterjee A, Aggarwal V, Ramos A, Acharya S, Thakor NV: A brain-computer interface with vibrotactile biofeedback for haptic information. J Neuroeng Rehabil 2007., 4: 10.1186/1743-0003-4-40
Liao X, Yao D, Li C: Transductive SVM for reducing the training effort in BCI. J Neural Eng 2007, 4: 246–254. 10.1088/1741-2560/4/3/010
Wei Q, Fei M, Wang Y, Gao X, Gao S: Feature combination for classifying single-trial ECoG during motor imagery of different sessions. Progress Natural Sci 2007, 17: 851–858. 10.1080/10002007088537482
Dornhege G, Blankertz B, Krauledat M, Losch F, Curio G, Mueller K-R: Combined optimization of spatial and temporal filters for improving brain-computer interfacing. IEEE Trans Biomed Eng 2006, 53: 2274–2281. 10.1109/TBME.2006.883649
Mahmoudi B, Erfanian A: Electro-encephalogram based brain-computer interface: improved performance by mental practice and concentration skills. Med Biol Eng Comput 2006, 44: 959–969. 10.1007/s11517-006-0111-8
Phothisonothai M, Nakagawa M: EEG-based classification of new imagery tasks using three-layer feedforward neural network classifier for brain-computer interface. J Physical Soc Japan 2006, 75: 104801. 10401–6 10.1143/JPSJ.75.104801
Ince NF, Arica S, Tewfik A: Classification of single trial motor imagery EEG recordings with subject adapted non-dyadic arbitrary time-frequency tilings. J Neural Eng 2006, 3: 235–244. 10.1088/1741-2560/3/3/006
Neuper C, Scherer R, Reiner M, Pfurtscheller G: Imagery of motor actions: differential effects of kinesthetic and visual-motor mode of imagery in single-trial EEG. Cogn Brain Res 2005, 25: 668–677. 10.1016/j.cogbrainres.2005.08.014
Burke DR, Kelly SR, de Chazal P, Reilly RB, Finucane C: A parametric feature extraction and classification strategy for brain-computer interfacing. IEEE Trans Neural Syst Rehabil Eng 2005, 13: 12–17. 10.1109/TNSRE.2004.841881
Kamousi B, Liu ZM, He B: Classification of motor imagery tasks for brain-computer interface applications by means of two equivalent dipoles analysis. IEEE Trans Neural Syst Rehabil Eng 2005, 13: 166–171. 10.1109/TNSRE.2005.847386
Kelly SP, Lalor EC, Reilly RB, Foxe JJ: Visual spatial attention tracking using high-density SSVEP data for independent brain-computer communication. IEEE Trans Neural Syst Rehabil Eng 2005, 13: 172–178. 10.1109/TNSRE.2005.847369
Treder MS, Schmidt NM, Blankertz B: Gaze-independent brain-computer interfaces based on covert attention and feature attention. J Neural Eng 2011., 8: 10.1088/1741-2560/8/6/066003
Chen M, Guan J, Liu H: Enabling fast brain-computer interaction by single-trial extraction of visual evoked potentials. J Med Syst 2011, 35: 1323–1331. 10.1007/s10916-011-9696-z
Kaufmann T, Schulz SM, Gruenzinger C, Kuebler A: Flashing characters with famous faces improves ERP-based brain-computer interface performance. J Neural Eng 2011., 8: 056016
Garcia Cossio E, Fernandez C, Eugenia Gaviria M, Palacio C, Alvaran L, Torres Villa RA: P300 based Brain computer interface for alternative communication: a case study with two teenagers with motor disabilities. Rev Fac Ing-Univ Antioquia 2011, 60: 9–19.
Kim DW, Hwang HJ, Lim JH, Lee YH, Jung KY, Im CH: Classification of selective attention to auditory stimuli: toward vision-free brain-computer interfacing. J Neurosci Methods 2011, 197: 180–185. 10.1016/j.jneumeth.2011.02.007
Pires G, Nunes U, Castelo-Branco M: Statistical spatial filtering for a P300-based BCI: Tests in able-bodied, and patients with cerebral palsy and amyotrophic lateral sclerosis. J Neurosci Methods 2011, 195: 270–281. 10.1016/j.jneumeth.2010.11.016
Mugler EM, Ruf CA, Halder S, Bensch M, Kubler A: Design and Implementation of a P300-Based Brain-Computer Interface for Controlling an Internet Browser. IEEE Trans Neural Syst Rehabil Eng 2011, 18: 599–609.
Lee PL, Sie JJ, Liu YJ, Wu CH, Lee MH, Shu CH, Li PH, Sun CW, Shyu KK: An SSVEP-actuated brain computer interface using phase-tagged flickering sequences: a cursor system. Ann Biomed Eng 2010, 38: 2383–2397. 10.1007/s10439-010-9964-y
Nam CS, Li YQ, Johnson S: Evaluation of P300-Based Brain-Computer Interface in Real-World Contexts. Int J Human-Comput Interact 2010, 26: 621–637. 10.1080/10447311003781326
Chen CW, Ju MS, Sun YN, Lin CCK: Model analyses of visual biofeedback training for EEG-based brain-computer interface. J Comput Neurosci 2009, 27: 357–368. 10.1007/s10827-009-0148-4
Bin G, Gao X, Yan Z, Hong B, Gao S: An online multi-channel SSVEP-based brain-computer interface using a canonical correlation analysis method. J Neural Eng 2009., 6: 046002
Klobassa DS, Vaughan TM, Brunner P, Schwartz NE, Wolpaw JR, Neuper C, Sellers EW: Toward a high-throughput auditory P300-based brain-computer interface. Clin Neurophysiol 2009, 120: 1252–1261. 10.1016/j.clinph.2009.04.019
Lee P-L, Hsieh J-C, Wu C-H, Shyu K-K, Wu Y-T: Brain computer interface using flash onset and offset visual evoked potentials. Clin Neurophysiol 2008, 119: 605–616. 10.1016/j.clinph.2007.11.013
Hoffmann U, Vesin J-M, Ebrahimi T, Diserens K: An efficient P300-based brain-computer interface for disabled subjects. J Neurosci Methods 2008, 167: 115–125. 10.1016/j.jneumeth.2007.03.005
Sellers EW, Krusienski DJ, McFarland DJ, Vaughan TM, Wolpaw JR: A P300 event-related potential brain-computer interface (BCI): The effects of matrix size and inter stimulus interval on performance. Biol Psychol 2006, 73: 242–252. 10.1016/j.biopsycho.2006.04.007
Lee P-L, Hsieh J-C, Wu C-H, Shyu K-K, Chen S-S, Yeh T-C, Wu Y-T: The brain computer interface using flash visual evoked potential and independent component analysis. Ann Biomed Eng 2006, 34: 1641–1654. 10.1007/s10439-006-9175-8
Lee PL, Wu CH, Hsieh JC, Wu YT: Visual evoked potential actuated brain computer interface: a brain-actuated cursor system. Electron Lett 2005, 41: 832–834. 10.1049/el:20050892
Lee P-L, Yeh C-L, Cheng JY-S, Yang C-Y, Lan G-Y: An SSVEP-Based BCI Using High Duty-Cycle Visual Flicker. IEEE Trans Biomed Eng 2011, 58: 3350–3359.
Volosyak I, Valbuena D, Luth T, Malechka T, Graser A: BCI Demographics II: How Many (and What Kinds of) People Can Use a High-Frequency SSVEP BCI? IEEE Trans Neural Syst Rehabil Eng 2011, 19: 232–239.
Takahashi H, Yoshikawa T, Furuhashi T: Error Control for Performance Improvement of Brain-Computer Interface: Reliability-Based Automatic Repeat Request. Ieice Transactions on Information and Systems 2011, E94D: 1243–1252.
Volosyak I: SSVEP-based Bremen-BCI interface-boosting information transfer rates. J Neural Eng 2011., 8: 036020
Wang YT, Wang YJ, Jung TP: A cell-phone-based brain-computer interface for communication in daily life. J Neural Eng 2011., 8: 025018
Blankertz B, Dornhege G, Krauledat M, Mueller K-R, Curio G: The non-invasive Berlin brain-computer interface: fast acquisition of effective performance in untrained subjects. Neuroimage 2007, 37: 539–550. 10.1016/j.neuroimage.2007.01.051
Royer AS, Rose ML, He B: Goal selection versus process control while learning to use a brain-computer interface. J Neural Eng 2011., 8: 036012
Vlek RJ, Schaefer RS, Gielen C, Farquhar JDR, Desain P: Sequenced subjective accents for brain-computer interfaces. J Neural Eng 2011., 8: 036002
Solis-Escalante T, Müller-Putz G, Brunner C, Kaiser V, Pfurtscheller G: Analysis of sensorimotor rhythms for the implementation of a brain switch for healthy subjects. Biomed Signal Process Control 2010, 5: 15–20. 10.1016/j.bspc.2009.09.002
Liu Y, Zhou ZT, Hu DW: Gaze independent brain-computer speller with covert visual search tasks. Clin Neurophysiol 2011, 122: 1127–1136. 10.1016/j.clinph.2010.10.049
Treder MS, Blankertz B: (C)overt attention and visual speller design in an ERP-based brain-computer interface. Behav Brain Funct 2010., 6: 10.1186/1744-9081-6-28
Nijboer F, Furdea A, Gunst I, Mellinger J, McFarland DJ, Birbaumer N, Kuebler A: An auditory brain-computer interface (BCI). J Neurosci Methods 2008, 167: 43–50. 10.1016/j.jneumeth.2007.02.009
Lee S, Lim H: Brain-Operated Typewriter using the Language Prediction Model. KSII Trans Internet Inf Syst 2011, 5: 1770–1782.
Zickler C, Riccio A, Leotta F, Hillian-Tress S, Halder S, Holz E, Staiger-Saelzer P, Hoogerwerf E-J, Desideri L, Mattia D, Kuebler A: A Brain-Computer Interface as Input Channel for a Standard Assistive Technology Software. Clin EEG Neurosci 2011, 42: 236–244. 10.1177/155005941104200409
Jutai JW, Fuhrer MJ, Demers L, Scherer MJ, DeRuyter F: Toward a Taxonomy of Assistive Technology Device Outcomes. Am J Phys Med Rehabil 2005, 84: 294–302. 10.1097/01.PHM.0000157313.88732.DC
Scherer MJ: Living in the State of Stuck: How Technology Impacts the Lives of People with Disabilities. Cambridge, MA: Brookline Books; 1993.
Scherer MJ: Outcomes of assistive technology use on quality of life. Disabil Rehabil 1996, 18: 439–448. 10.3109/09638289609165907
Sellers EW, Donchin E: A P300-based brain-computer interface: Initial tests by ALS patients. Clin Neurophysiol 2006, 117: 538–548. 10.1016/j.clinph.2005.06.027
Wolpaw JR: An EEG-based brain-computer interface for cursor control. Electroencephalogr Clin Neurophysiol 1991, 78: 252–259. 10.1016/0013-4694(91)90040-B
Gao Y, Black MH: A quantitative comparison of linear and non-linear models of motor cortical activity for the encoding and decoding of arm motions. Int IEEE EMBS Conf Neural Eng 2003, 189–192. 10.1109/CNE.2003.1196789
Nykopp T: Statistical modelling issues for the adaptive brain interface. Helsinki, Finland: Helsinki University of Technology; 2001.
Blankertz B, Muller K, Curio G, Vaughan TM, Schalk G, Wolpaw J, Schlogl A, Neuper C, Pfurtscheller G, Hinterberger T, Schroder M, Birbaumer N: The BCI Competition 2003: progress and perspectives in detection and discrimination of EEG single trials. IEEE Trans Biomed Eng 2003, 2004(51):1044–1052.
Cohen J: A coefficient of agreement for nominal scales. Educ Psychol Meas 1960, 20: 37–46. 10.1177/001316446002000104
Schlogl A, Lee FY: Characterization of Four-Class Motor Imagery EEG Data for the BCI-Competition 2005. J Neural Eng 2005, 2: 14–22. 10.1088/1741-2560/2/4/L02
Gwet K: Inter-rater reliability: Dependency on trait prevalence and marginal homogeneity. Statistical Methods for Inter-Rater Reliability Assessment Series 2002, 2: 1–9.
Obermaier B, Neuper C, Guger C, Pfurtscheller G: Information transfer rate in a five-classes brain-computer interface. IEEE Trans Neural Syst Rehabil Eng 2001, 9: 283–288. 10.1109/7333.948456
Schlogl A, Neuper C, Pfurtscheller G: Estimating the mutual information of an EEG-based brain-computer interface. Biomed Tech 2002, 47: 3–8.
Wolpaw J, Ramoser H: EEG-based communication: improved accuracy by response verification. IEEE Trans Rehabil Eng 1998, 6: 326–333. 10.1109/86.712231
Kronegg J, Voloshynovskiy S, Pun T: Analysis of bit-rate definitions for Brain-Computer Interfaces. Las Vegas: Conference on Human-Computer Interactions; 2005.
Buttfield A, Ferrez PW, Millan JR: Towards a robust BCI: Error potentials and online learning. IEEE Trans Neural Syst Rehabil Eng 2006, 14: 164–168. 10.1109/TNSRE.2006.875555
Baker B: Minspeak: A semantic compaction system that makes self-expression easier for communicatively disabled individuals. Byte 1982, 7: 186–202.
Zhang D, Wang Y, Gao X, Hong B, Gao S: An algorithm for idle-state detection in motor-imagery-based brain-computer interface. Comput Intell Neurosci 2007. 39714
Thompson DE, Gruis KL, Huggins JE: A plug-and-play brain-computer interface to operate commercial assistive technology. Disabil Rehabil Assist Technol 2013. in press PMID: 23590556
Mason SG, Kronegg J, Huggins J, Fatourechi M, Schlogl A: Evaluating the Performance of Self-Paced Brain-Computer Interface Technology. 2006. http://ipl.ece.ubc.ca/bci_files/self_paced_tech_report-2006–05–19.pdf
Panicker RC, Puthusserypady S: Ying Sun: An Asynchronous P300 BCI With SSVEP-Based Control State Detection. IEEE Trans Biomed Eng 2011, 58: 1781–1788.
Davis J, Goadrich M: The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd international conference on Machine learning. New York, NY, USA: ACM; 2006:233–240. [ICML’06]
Schogl A, Anderer P: Artefact detection in sleep EEG by the use of Kalman filtering. Vienna, Austria: Proceedings EMBEC; 1999. [EMBEC]
Simpson RC, Koesten HH: Adaptive one-switch row-column scanning. IEEE Trans Rehabil Eng 1999, 7: 464–473. 10.1109/86.808950
This work was supported in part by the National Institute of Child Health and Human Development (NICHD), the National Institutes of Health (NIH) under Grant R21HD054697 and by the National Institute on Disability and Rehabilitation Research (NIDRR) in the Department of Education under Grant H133G090005 and Award Number H133P090008. This material also is based upon work supported by the National Science Foundation Graduate Student Research Fellowship under Grant DGE0718128. Any opinions, findings, conclusions or recommendations expressed in this publication are those of the authors and do not necessarily reflect the views of the NICHD, NIH, NIDRR, the Department of Education, or the NSF.
The authors declare that they have no competing interests.
DET performed the mathematical analysis, summarized and rated each metric. He also developed the criteria jointly with SBM. SBM performed the literature review, developed the conceptual framework, and prepared the manuscript. JEH participated in the discussion of metrics, helping to shape the selection criteria and select metrics for detailed analysis. All authors read and approved the final manuscript.
Stefanie Blain-Moraes contributed equally to this work.
About this article
Cite this article
Thompson, D.E., Blain-Moraes, S. & Huggins, J.E. Performance assessment in brain-computer interface-based augmentative and alternative communication. BioMed Eng OnLine 12, 43 (2013). https://doi.org/10.1186/1475-925X-12-43