- 1Department of Psychology, University of New Mexico, Albuquerque, NM, United States
- 2Department of Computer Science, University of New Mexico, Albuquerque, NM, United States
Electroencephalographic (EEG) recordings are thought to reflect the network-wide operations of canonical neural computations, making them a uniquely insightful measure of brain function. As evidence of these virtues, numerous candidate biomarkers of different psychiatric and neurological diseases have been advanced. Presumably, we would only need to apply powerful machine-learning methods to validate these ideas and provide novel clinical tools. Yet, the reality of this advancement is more complex: the scale of data required for robust and reliable identification of a clinical biomarker transcends the ability of any single laboratory. To surmount this logistical hurdle, collective action and transparent methods are required. Here we introduce the Patient Repository of EEG Data + Computational Tools (PRED+CT: predictsite.com). The ultimate goal of this project is to host a multitude of available tasks, patient datasets, and analytic tools, facilitating large-scale data mining. We hope that successful completion of this aim will lead to the development of novel EEG biomarkers for differentiating populations of neurological and psychiatric disorders.
Introduction
There is a critical need to standardize and quantify the diagnostic criteria for psychiatric and neurological disorders. A lack of uniform diagnostic schema and reliance on phenotypic assessment has resulted in critical gaps in clinical practice between institutions. A growing number of reports suggest that biomarkers or endophenotypes (endogenous phenotypes) will be more effective for diagnosis and disease classification than an expanded phenotypic characterization (Gould and Gottesman, 2006; Diaz-Arrastia et al., 2009; Robbins et al., 2012).
Our long-term goal is to develop and maintain an open-source website that leverages the power of collective action to address this need using electroencephalography (EEG). By quantifying the emergence of psychological operations at their source, EEG provides a mechanistic means to transcend the descriptive, correlative, and phenotypic descriptions of symptomatology that currently characterizes this field. EEG is less expensive, more readily available, highly portable, and simpler to operate than competing imaging recourses, making it logistically viable.
While there are a large number of open-source EEG data repositories, these are sparsely populated (Table 1). Patient-specific online repositories tend not to have EEG data. While some sites may contain both EEG data and patient groups, they sometimes require formal requests and selective processes for data acquisition, and tend to not include matched controls. The absence of open-source software infrastructure for standardized and large-scale EEG data reflects a failure of basic biomedical planning and initiative. We aim to fill this gap with a one-stop open-source site for gathering, storing, and analyzing clinically relevant data. In this report we present the Patient Repository of EEG Data + Computational Tools (PRED+CT1).
Current and Future Use of EEG As A Biomarker
Electroencephalography-based biomarkers are particularly salient due to current widespread use in neurology clinics, which increases the likelihood that a novel advancement will have immediate clinical significance. Some neural deficits like epilepsy are objectively diagnosable following an EEG; other complications like tumors and stroke can be inferred. Yet disorders that affect higher cognitive functions remain opaque following any type of routine imaging. The future use of EEG as a clinical biomarker aims to capitalize on this existing diagnostic infrastructure via knowledge advancements, facilitating greater diagnostic utility from already routine scanning sessions.
Electroencephalography is uniquely sensitive to canonical neural operations which underlie emergent psychological constructs (Fries, 2009; Siegel et al., 2012; Cavanagh and Castellanos, 2016), making it well suited for discovery of aberrant neural mechanisms that underlie complicated disease states (Insel et al., 2010; Montague et al., 2012). As an example, consider how error-related EEG activities can sensitively and specifically dissociate generalized anxiety participants from healthy controls (Cavanagh et al., 2017). This finding follows positive reports from two meta-analyses with 37 and 46 studies, and 1757 and 1616 participants, respectively (Moser et al., 2013; Cavanagh and Shackman, 2014). If even a fraction of the studies included in these meta-analyses were openly available, a widely generalizable set of discriminating features would already be available for extended research and possible clinical application.
The Need for An Online Repository Dedicated to this Purpose
Findings in cognitive neuroscience tend to advance through independent laboratories working separately, with each group developing their own stimulus presentation tasks and data analysis parameters. While this flexibility is beneficial, it is also a threat to external validity. Generalizability is critical to consider since the scale of data required to effectively characterize clinical biomarkers transcends the abilities of any single laboratory. To realize this goal of differential diagnostic biomarkers, open-source collaboration across institutions will be required. Currently, the field of EEG lacks even a rudimentary foundation for this goal.
The need for open-source data sharing and a focus on replicability has been widely approached in the MRI community, with varied types of data repositories (Gorgolewski et al., 2015; Eickhoff et al., 2016; Iyengar, 2016; Nichols et al., 2017; Poldrack et al., 2017), including patient-specific databases and depictions of classification goals (Woo et al., 2017). PRED+CT uses the OpenfMRI project (Poldrack et al., 2013) as a model. PRED+CT will not only be the first open-source EEG database for patient data, but it will work to standardize assessment and analytic tools, facilitating the overarching goal of distributed data collection and data mining.
Prior Approaches
It is important to note that although this basic clinical goal has been addressed with these basic techniques for a long time, the approach we advance here offers a significant advancement from the status quo. While visual inspection of EEG is still the normative procedure in neurology, the clear potential of computer-based assessments spawned a more quantitative approach over a generation ago.
Quantitative EEG (QEEG) summarizes an approach to EEG assessment on a few minutes of artifact-free data in a resting state, usually on a 19-channel clinical setup (John et al., 1988; Nuwer, 1997; Prichep, 2005; Coburn et al., 2006). The most common QEEG approaches use fast Fourier transforms (FFTs) to compute absolute and relative power at sites, hemispheric asymmetry of power, power ratios between pre-defined frequency bands, squared correlation of activity (“coherence”), phase lag times between electrode sites, or other related measures. These features are used to populate large-scale normative datasets to contrast with patient-specific databases, oftentimes statistically controlling for spurious variables like age. Finally, classification procedures like cross-validation and algorithms like discriminant analysis or neural networks may be applied in order to identify features that maximally discriminate patients from controls. This approach rests on the thesis that reliable statistical differentiation in the spatial representations of these multidimensional activities can differentiate a wide variety of psychiatric, congenital, and neurological disorders.
Notable successes include a Food and Drug Administration-approved prognostic biomarker for Attention-Deficit Hyperactivity Disorder in the ratio of theta to beta band power at the vertex electrode (Food and Drug Administration, 2013), and a candidate biomarker for acute traumatic brain injury (Thatcher et al., 1989; Naunheim et al., 2010; Prichep et al., 2014). While the validity and appropriate clinical utilization of these procedures remain highly debated (Arns et al., 2013; Saad et al., 2015; Gloss et al., 2016), resolution of these issues may be hamstrung by questionable premises underlying the general practice of QEEG (described below). We believe that the analytic approach motivated by PRED+CT will successfully address these problems and facilitate significant advancement in this field.
This QEEG approach has been highly controversial for a long time for a large number of reasons (Nuwer, 1997; Kaiser, 2000; Thatcher et al., 2003; Nuwer et al., 2005; Coburn et al., 2006). First, simple dissociation from statistical normality can easily be caused by the influence of spurious variables. Statistical dissociation has neither face-valid clinical implication nor clinical utility, as it doesn’t demonstrate that the differentiation will be faithfully reflected at the level of an individual (Amyot et al., 2015). Second, the QEEG approach relies primarily on resting activity, which has highly varied reliability across derived features (Nuwer et al., 2005), and lacks content and construct validity for assessing psychiatric and neurological disorders. Diagnostic criteria may even be the wrong target for associating with brain scans: identification of aberrant neural mechanisms underlying a disorder may be a more fruitful target than phenotypic characterization (Insel et al., 2010; Montague et al., 2012; Gillan and Daw, 2016). Third, the overleveraging of the FFT offers only a superficial decomposition of brain activities. Few studies have aimed to apply highly novel analytic techniques to derive maximally dissociating features of specific diseases (cf. Allen and Cohen, 2010; Lainscsek et al., 2013). Together, these issues reflect a fundamental failure to appreciate how EEG activities mechanistically reflect unique neural computations that may most parsimoniously define disease states.
In some cases, black-box techniques for pre-processing and quantification have been intertwined with privatization and commercialization. While commercialization is an important positive step toward clinical utility, it is sometimes necessarily closed source and has unfortunately been associated with overzealous advertisement and dubious claims (Nuwer et al., 2005; Coburn et al., 2006). In sum, while QEEG is a promising and selectively successful approach, current practice is highly limited. PRED+CT aims to motivate a common platform for methodological advancement, predicated on fully transparent databases and computational tools. Importantly, bigger data and increased algorithmic complexity are only part of the solution: we would like to highlight that success may depend on the ability to identify the right task to probe the aberrant mechanism underlying a specific disorder.
PRED+CT
The ultimate goal of this project is to host a multitude of tasks and patient datasets, facilitating large-scale data mining (Figure 1). We hope that successful completion of this aim will lead to development of novel EEG biomarkers with enhanced predictive power above and beyond phenotypic assessments for differentiating populations of neurological and psychiatric disorders.
FIGURE 1. Screenshot of the PRED+CT home screen (www.predictsite.com).
Tasks
To facilitate standardization across laboratories, we have developed software applications for oddball-type tasks (Sutton et al., 1965) and an Eriksen flankers task (Eriksen and Eriksen, 1974). These applications are coded in the Java programming language and work on Windows, but can also be run on other platforms through a Windows emulator. Each application allows the user to input a range of configurations to comprehensively cover parameter variations (inter-trial interval, visual or auditory modality, error feedback, instructions, etc.). A user can save a configuration file with these parameters, as well as any type of response requirement. This latter feature facilitates the creation of a variety of passive and active tasks, including go/no-go and vigilance tasks. By providing these programs, we hope to encourage smaller site-specific patient studies to include an additional short assessment to their protocols for the purpose of open-source data sharing.
If popular, we hope to include a multitude of task types in the future, such as reward gambling, stop-signal, basic language, and motor tasks, etc. Some existing task batteries for EEG assessment capitalize on simultaneous acquisition of a large number of EEG events/ERP components (Kappenman and Luck, 2012; Kieffaber et al., 2016; Nair et al., 2016); this may be a promising direction for future expansion.
Upload and Download
An upload tab facilitates user requests for contributing data to PRED+CT. A download tab contains study information (Table 2) and will be fully open (no log in or request required). All data will be hosted in Matlab readable format (.mat or .set files, which are interchangeable) for a few reasons. Matlab is the current most common platform for academic EEG research, and the popular EEGLab suite (Delorme and Makeig, 2004) will be utilized as a common data structure. Many native file types contain information that could be a threat to confidentiality, and EEGLab import tools strip many of these markers. Matlab files are easily imported into other (open-source) programs like Octave, Python, and R, so this common structure should not be limiting in any way.
TABLE 2. Information required for datasets (EEG files) and tools (computer programs for analysis) to be contributed to PRED+CT.
Confidentiality
The most likely threats to personal health information in candidate PRED+CT database entries are names or initials, locations, and dates. The user is required to ensure that none of these remain in the subject identifier or in the EEG metadata: for example, BrainVision.vhdr files contain times and .vmrk files contain a date stamp. EEGLab import to Matlab strips the data of such possible hidden threats to confidentiality. PRED+CT administrators will perform a double check on potential threats to confidentiality and can assist with the translation from native formats into .mat files. Most institutional review boards (IRBs) do not consider data sharing itself to fall under the definition of “human subjects research,” but interested users should request a determination from their IRB. Including such language in the informed consent is the best way to ensure that ethical issues are proactively well-managed.
Data Organization
It is highly recommended that users include data in its raw form prior to any pre-processing; however, this may be infeasible in some cases. In many instances, important recording information is embedded in the native data format (sampling rate, reference, electrode labels) and is translated directly into the EEG data structure. If this information is missing (i.e., data are from a clinical system) then the user should include a readme file with this information.
For data other than rest, is important that users can understand what trigger types (TTL pulses) are used to represent each type of event. At minimum, a comprehensive list of trigger types needs to be included, and we highly recommend that users include the stimulus presentation script (i.e., Eprime run file, Matlab Psychtoolbox files, etc.). Ideally, the task was designed so that behavioral responses can be recoverable from the triggers; if not then upload of separate behavioral logs is encouraged. While there are common neuroinformatics structures that can facilitate sophisticated data organization schemes across studies (Teeters et al., 2008; Landis et al., 2016; Plis et al., 2016; Wiener et al., 2016), we opted for a more simplified approach in PRED+CT based on simple and well-documented descriptions of idiosyncratic TTL triggers.
Computational Tools
Pattern classifiers use cross-validation or bootstrap approaches where the whole dataset is partitioned into non-overlapping training and test sets. The classifiers involve optimizing information theoretic and multidimensional metrics to generate models based on signal shapes and do not over-fit the data as traditional predictive models do (Parra et al., 2005; Pereira et al., 2009; Lemm et al., 2011). Such pattern-based classification is thus necessary to generalize predictive models for diagnostic subtyping and recovery trajectory to other groups (aka biomarkers).
There are a number of tutorials for general brain science classification (Pereira et al., 2009; Lemm et al., 2011), EEG-specific tutorials (Parra et al., 2005; Dyrholm and Parra, 2006; Schirrmeister et al., 2017), and open-source sets of analytic tools (Detre et al., 2006; Hanke et al., 2009). Our goal is not to replicate these resources, but to provide a repository for computational approaches that can bolster feature selection or patient classification, particularly on existing datasets in the archive. While we envision hosting a multitude of scripts, there is no reason that entries on this page couldn’t link to external resources (i.e., github or NITRC).
Intellectual Property and Credit
Unless otherwise noted, this database and its contents are made available under the Public Domain Dedication and License v1.0 whose full text can be found at: https://opendatacommons.org/licenses/pddl/1.0/. We hope that all users will follow the ODC Attribution/Share-Alike Community Norms, including the expectation that there will be no attempt to de-anonymize any data.
Example
Figure 2 shows an example of the type of outcome we hope to cultivate with PRED+CT. This figure shows a receiver operating characteristic plot detailing classification of Parkinson’s patients on and off medication vs. well-matched controls based on three-auditory oddball task conditions (Cavanagh et al., under review). That report details the reasons why aberrant orienting to novelty is mechanistically interesting in Parkinson’s disease, and why EEG is uniquely well-suited to assess to biomarker potential of the associated neural response. These raw data and scripts are available on the PRED+CT website (accession nos.: d001 and t001). Since this task is very brief and very easy for patients to perform, we encourage other groups to contribute similar datasets to examine the generalizability of this phenomenon.
FIGURE 2. Example support vector machine (SVM) classification of Parkinson’s patients ON and OFF medication vs. well-matched controls based on three-auditory oddball task conditions. (A) A receiver operating characteristic plot shows the true vs. false positive rates of PD vs. CTL discrimination for each medication and task condition. (B) Total accuracy (average of sensitivity and specificity) for each condition. (C) Correlation of SVM confidence and years diagnosed for each condition. EEG data are available under Downloads (accession no.: d001) and Matlab scripts are available under Computational Tools (accession no.: t001).
Updates
Interested users can follow @PREDiCT_Admin or #PREDiCT + #UNM on Twitter for updates, including new dataset and tool contributions.
Future Challenges to Surmount
Challenges in Data Processing
While data sharing is “good,” a prevalent challenge is to share high-quality usable data (Kennedy, 2004). By archiving EEG and metadata in a common EEGLab structure, we can fulfill these criteria. Substantive hardware and software advancements over time are unlikely to change basic aspects of EEG data. Numerous algorithms exist to assist in pre-processing EEG data (Delorme and Makeig, 2004; Nolan et al., 2010; Oostenveld et al., 2011; Bigdely-Shamlo et al., 2015; Chaumon et al., 2015). In the Computational Tools section, we have provided our Algorithmic Pre-Processing Line for EEG (APPLE.m; accession no.: t002), which leverages a combination of FASTER (Nolan et al., 2010), ADJUST (Mognon et al., 2011), EEGLab, and custom algorithms for automatically interpolating bad channels, removing bad epochs, and identifying the most likely independent component associated with eye blinks.
Electroencephalography datasets come in a variety of reference schemes, topographical layouts, and sampling rates, complicating integration. However, these are addressable problems. Use of the average reference and relative measurements (decibel, percent change, relative power) facilitate common analytic space. EEG data tends to be oversampled, so down-sampling to a common lowest denominator is a viable option. As pattern classifiers leverage any difference between training sets, it will be critical to ensure that spurious differences between combined datasets do not interfere with the aim of classifying patients from controls. Having an equal number of patients and well-matched controls in each dataset is a good first step for experimental control over this issue, but additional steps like controlled randomization of training and testing sets may be required to control for dataset-specific biases.
Once pre-processed, EEG data offer a rather simple data structure that is accessible by non-experts. A two-dimensional matrix of channels ∗ time can be easily restructured to include a third dimension based on discrete events (i.e., the EEG.data field of EEGlab), and no special software, opaque statistical constraints, advanced processing, or other complicated considerations are strictly necessary for interpretation. We hope this increases the appeal to computer and data scientists, who should be able to manage EEG data as an input variable with very minimal special training.
Challenges in Prediction
A well-known adage in machine learning is that achieving 80% classification accuracy is easy, and closing the gap toward 100% accuracy will take between a few years and eternity. We think that PRED+CT can assist with strategies for (partially) closing this gap, which we detail in order of their intuitiveness. The most straightforward solution to boost generalizability is to utilize larger training sets (i.e., more EEG data), which is the primary purpose of the site. Another immediately apparent solution is to leverage algorithmic advancements. In isolation it is hard to know how revolutionary different classification procedures will be, but open communication may help set standards and constraints on parameter selection which otherwise act as a hidden threat to generalizability. It is important to note that theoretical validation requires the ability to interpret which of the input features led to successful classification (cf. Steele et al., 2014; Cavanagh and Castellanos, 2016; Doshi-Velez and Kim, 2017).
We submit that the best way to achieve these overarching goals will involve taking advantage of the EEG feature(s) reflecting the neural computations that maximally discriminate groups. While resting activities may be the optimal solution for some patient groups, specifically designed active tasks are likely necessary to elicit the requisite brain responses that characterize the nature of the departure from statistical normality. Error signaling in anxiety has already been described above as a defining neural computation related to the etiology of the disorder, but other candidate responses have been advanced for other disorders, including diminished reward signals for major depression (Proudfit, 2015), a broken target-updating P3b in schizophrenia (Ford, 1999), a reduced novelty orienting P3a in Parkinson’s (Solís-Vivanco et al., 2015), and reduced brainstem evoked responses in acute traumatic brain injury (Kraus et al., 2016) to name just a few.
Finally, a less intuitive recipe for success may be to simply ask more specific questions. More constrained hypotheses can help to collapse insurmountable prior probabilities into the realm of plausibility. Consider that the base rate for any specific neurological or psychiatric disease is low enough to dismiss the plausibility of a new EEG-based diagnostic test with viable sensitivity and specificity. Yet if a patient is already being treated for a disease, this obviates some base rate problems. For example, instead of trying to develop a fast and easy brain scan to identify if someone has major depressive disorder, it is more plausible to ask if a diminished brain response to reward can help guide treatment options to address melancholic vs. atypical features of depression.
Challenges in Diagnostics
To be medically useful, a test must have positive prognostic value above and beyond current status quo, or reduce time, cost, or uncertainty (Nuwer et al., 2005). While brain-based diagnostics may achieve impressive sensitivity, they are often associated with high false positives (low specificity), which is a particular deterrent to clinical use for differential diagnosis (Nuwer et al., 2005). These challenges are addressable, especially since direct clinical application is not a necessary outcome of brain-based patient classification.
High sensitivity in the context of low specificity may nevertheless have important clinical utility for rapidly assessing the potential presence of diagnostic complications (Hanley et al., 2013; Ayaz et al., 2015), quantifying the degree of injury severity (Thatcher et al., 2001), or for tracking differences in disease progression in treatment studies. Identification of the maximally discriminable neural computation that defines a patient group has additional translational utility: for instance it could be used as concurrent validation of a novel imaging or biomarker measure. In sum, reliable novel findings are important successes even if they do not lead to direct clinical translation.
Conclusion
The genesis of PRED+CT was motivated by an understanding of the strength of EEG measurements and methods, but equally matched frustration in the logistical constraints of advancing beyond small-scale validation studies. EEG is a uniquely powerful measure of canonical neural operations, and machine learning has already led to profound social advancements. Surely we should have some firm answers to important clinical neuroscience questions by now. Yet single laboratory contributions to clinical science remain slow, expensive, time-consuming, and oftentimes led to beautiful but neglected datasets as interests and energies are applied to new funding opportunities. Only through collective action and full transparency can we hope to realize the utility of EEG-derived features of underlying neural computations for clinical neuroscience research.
Author Contributions
JC and AM designed the project. AN and CW programed the site and tools. JC wrote the first draft.
Funding
This work was supported by a grant from the University of New Mexico Office of the Vice President of Research. JC and AM are also supported by NIGMS 1P20GM109089-01A1.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
The authors thank Lynette Bustos for the artwork that adorns the site. PRED+CT can be accessed at www.predictsite.com.
Footnotes
References
Allen, J. J. B., and Cohen, M. X. (2010). Deconstructing the “resting” state: exploring the temporal dynamics of frontal alpha asymmetry as an endophenotype for depression. Front. Hum. Neurosci. 4:232. doi: 10.3389/fnhum.2010.00232
Amyot, F., Arciniegas, D. B., Brazaitis, M. P., Curley, K. C., Diaz-Arrastia, R., Gandjbakhche, A., et al. (2015). A review of the effectiveness of neuroimaging modalities for the detection of traumatic brain injury. J. Neurotrauma 32, 1693–1721. doi: 10.1089/neu.2013.3306
Arns, M., Conners, C. K., and Kraemer, H. C. (2013). A decade of EEG theta/beta ratio research in ADHD: a meta-analysis. J. Atten. Disord. 17, 374–383. doi: 10.1177/1087054712460087
Ayaz, S. I., Thomas, C., Kulek, A., Tolomello, R., Mika, V., Robinson, D., et al. (2015). Comparison of quantitative EEG to current clinical decision rules for head CT use in acute mild traumatic brain injury in the ED. Am. J. Emerg. Med. 33, 493–496. doi: 10.1016/j.ajem.2014.11.015
Bigdely-Shamlo, N., Mullen, T., Kothe, C., Su, K.-M., and Robbins, K. A. (2015). The PREP pipeline: standardized preprocessing for large-scale EEG analysis. Front. Neuroinform. 9:16. doi: 10.3389/fninf.2015.00016
Cavanagh, J. F., and Castellanos, J. (2016). Identification of canonical neural events during continuous gameplay of an 8-bit style video game. Neuroimage 133, 1–13. doi: 10.1016/j.neuroimage.2016.02.075
Cavanagh, J. F., Meyer, A., Hajcak, G., Schroder, H. S., Larson, M. J., Jonides, J., et al. (2017). Error-specific cognitive control alterations in generalized anxiety disorder. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 53, 21–29. doi: 10.1016/j.bpsc.2017.01.004
Cavanagh, J. F., and Shackman, A. J. (2014). Frontal midline theta reflects anxiety and cognitive control: meta-analytic evidence. J. Physiol. Paris 109, 3–15. doi: 10.1016/j.jphysparis.2014.04.003
Chaumon, M., Bishop, D. V. M., and Busch, N. A. (2015). A practical guide to the selection of independent components of the electroencephalogram for artifact correction. J. Neurosci. Methods 250, 47–63. doi: 10.1016/j.jneumeth.2015.02.025
Coburn, K. L., Lauterbach, E. C., Boutros, N. N., Black, K. J., Arciniegas, D. B., and Coffey, C. E. (2006). The value of quantitative electroencephalography in clinical psychiatry: a report by the committee on research of the American neuropsychiatric association. J. Neuropsychiatr. 18, 460–500. doi: 10.1176/appi.neuropsych.18.4.460
Delorme, A., and Makeig, S. (2004). EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134, 9–21. doi: 10.1016/j.jneumeth.2003.10.009S0165027003003479
Detre, G., Polyn, S., Moore, C., and Natu, V. (2006). The Multi-Voxel Pattern Analysis (MVPA) Toolbox. Available at: http://scholar.google.com/scholar?hl=en&btnG=Search&q=intitle:The+Multi-Voxel+Pattern+Analysis+(MVPA)+Toolbox#0
Diaz-Arrastia, R., Agostini, M. A., Madden, C. J., and Van Ness, P. C. (2009). Posttraumatic epilepsy: the endophenotypes of a human model of epileptogenesis. Epilepsia 50(Suppl. 2), 14–20. doi: 10.1111/j.1528-1167.2008.02006.x
Doshi-Velez, F., and Kim, B. (2017). Towards a Rigorous Science of Interpretable Machine Learning. Available at: http://arxiv.org/abs/1702.08608
Dyrholm, M., and Parra, L. C. (2006). Smooth bilinear classification of EEG. Annu. Int. Conf. IEEE Eng. Med. Biol. Proc. 1, 4249–4252. doi: 10.1109/IEMBS.2006.260083
Eickhoff, S., Nichols, T. E., Van Horn, J. D., and Turner, J. A. (2016). Sharing the wealth: neuroimaging data repositories. Neuroimage 124, 1065–1068. doi: 10.1016/j.neuroimage.2015.10.079
Eriksen, B. A., and Eriksen, C. (1974). Effects of noise letters upon the identification of a target letter in a nonsearch task. Percept. Psychophys. 16, 143–149. doi: 10.3758/BF03203267
Food and Drug Administration (2013). De Novo Classification Request for Neuropsychiatric EEG-Based Assessment Aid for ADHD (NEBA) System. Available at: https://www.accessdata.fda.gov/cdrh_docs/reviews/K112711.pdf
Ford, J. (1999). Schizophrenia: the broken P300 and beyond. Psychophysiology 36, 667–682. doi: 10.1111/1469-8986.3660667
Fries, P. (2009). Neuronal gamma-band synchronization as a fundamental process in cortical computation. Annu. Rev. Neurosci. 32, 209–224. doi: 10.1146/annurev.neuro.051508.135603
Gillan, C. M., and Daw, N. D. (2016). Taking psychiatry research online. Neuron 91, 19–23. doi: 10.1016/j.neuron.2016.06.002
Gloss, D., Varma, J. K., Pringsheim, T., and Nuwer, M. R. (2016). Practice advisory: the utility of EEG theta/beta power ratio in ADHD diagnosis. Neurology 87, 2375–2379. doi: 10.1212/WNL.0000000000003265
Gorgolewski, K. J., Varoquaux, G., Rivera, G., Schwarz, Y., Ghosh, S. S., Maumet, C., et al. (2015). NeuroVault.org: a web-based repository for collecting and sharing unthresholded statistical maps of the human brain. Front. Neuroinform. 9:8. doi: 10.3389/fninf.2015.00008
Gould, T. D., and Gottesman, I. I. (2006). Psychiatric endophenotypes and the development of valid animal models. Genes. Brain. Behav. 5, 113–119. doi: 10.1111/j.1601-183X.2005.00186.x
Hanke, M., Halchenko, Y. O., Sederberg, P. B., Hanson, S. J., Haxby, J. V., and Pollmann, S. (2009). PyMVPA: a python toolbox for multivariate pattern analysis of fMRI data. Neuroinformatics 7, 37–53. doi: 10.1007/s12021-008-9041-y
Hanley, D. F., Chabot, R., Mould, W. A., Morgan, T., Naunheim, R., Sheth, K. N., et al. (2013). Use of brain electrical activity for the identification of hematomas in mild traumatic brain injury. J. Neurotrauma 30, 2051–2056. doi: 10.1089/neu.2013.3062
Insel, T., Cuthbert, B., Garvey, M., Heinssen, R., Pine, D. S., Quinn, K., et al. (2010). Research domain criteria (RDoC): toward a new classification framework for research on mental disorders. Am. J. Psychiatry 167, 748–751. doi: 10.1176/appi.ajp.2010.09091379
Iyengar, S. (2016). Case for fMRI data repositories. Proc. Natl. Acad. Sci. U.S.A. 113, 7699–7700. doi: 10.1073/pnas.1608146113
John, E. R., Prichep, L. S., Fridman, J., and Easton, P. (1988). Neurometrics: computer-assisted differential diagnosis of brain dysfunctions. Science 239, 162–169. doi: 10.1126/science.3336779
Kaiser, D. A. (2000). QEEG: state of the art, or state of confusion. J. Neurother. 4, 57–75. doi: 10.1300/J184v04n02_07
Kappenman, E. S., and Luck, S. J. (2012). Manipulation of orthogonal neural systems together in electrophysiological recordings: the MONSTER approach to simultaneous assessment of multiple neurocognitive dimensions. Schizophr. Bull. 38, 92–102. doi: 10.1093/schbul/sbr147
Kennedy, D. N. (2004). Barriers to the socialization of information. Neuroinformatics 2, 367–368. doi: 10.1385/NI:2:4:367
Kieffaber, P. D., Okhravi, H. R., Hershaw, J. N., and Cunningham, E. C. (2016). Evaluation of a clinically practical, ERP-based neurometric battery: application to age-related changes in brain function. Clin. Neurophysiol. 127, 2192–2199. doi: 10.1016/j.clinph.2016.01.023
Kraus, N., Thompson, E. C., Krizman, J., Cook, K., White-Schwoch, T., and LaBella, C. R. (2016). Auditory biological marker of concussion in children. Sci. Rep. 6:39009. doi: 10.1038/srep39009
Lainscsek, C., Hernandez, M. E., Weyhenmeyer, J., Sejnowski, T. J., and Poizner, H. (2013). Non-linear dynamical analysis of EEG time series distinguishes patients with Parkinson’s disease from healthy individuals. Front. Neurol. 4:200. doi: 10.3389/fneur.2013.00200
Landis, D., Courtney, W., Dieringer, C., Kelly, R., King, M., Miller, B., et al. (2016). COINS data exchange: an open platform for compiling, curating, and disseminating neuroimaging data. Neuroimage 124, 1084–1088. doi: 10.1016/j.neuroimage.2015.05.049
Lemm, S., Blankertz, B., Dickhaus, T., and Müller, K.-R. (2011). Introduction to machine learning for brain imaging. Neuroimage 56, 387–399. doi: 10.1016/j.neuroimage.2010.11.004
Mognon, A., Jovicich, J., Bruzzone, L., and Buiatti, M. (2011). ADJUST: An automatic EEG artifact detector based on the joint use of spatial and temporal features. Psychophysiology 48, 229–240. doi: 10.1111/j.1469-8986.2010.01061.x
Montague, P. R., Dolan, R. J., Friston, K. J., and Dayan, P. (2012). Computational psychiatry. Trends Cogn. Sci. 16, 72–80. doi: 10.1016/j.tics.2011.11.018
Moser, J. S., Moran, T. P., Schroder, H. S., Donnellan, M. B., and Yeung, N. (2013). On the relationship between anxiety and error monitoring: a meta-analysis and conceptual framework. Front. Hum. Neurosci. 7:466. doi: 10.3389/fnhum.2013.00466
Nair, A. K., Sasidharan, A., John, J. P., Mehrotra, S., and Kutty, B. M. (2016). Assessing neurocognition via gamified experimental logic: a novel approach to simultaneous acquisition of multiple ERPs. Front. Neurosci. 10:1. doi: 10.3389/fnins.2016.00001
Naunheim, R. S., Treaster, M., English, J., Casner, T., and Chabot, R. (2010). Use of brain electrical activity to quantify traumatic brain injury in the emergency department. Brain Inj. 24, 1324–1329. doi: 10.3109/02699052.2010.506862
Nichols, T. E., Das, S., Eickhoff, S. B., Evans, A. C., Glatard, T., Hanke, M., et al. (2017). Best practices in data analysis and sharing in neuroimaging using MRI. 20, 299–303. doi: 10.1038/nn.4500
Nolan, H., Whelan, R., and Reilly, R. B. (2010). FASTER: fully automated statistical thresholding for EEG artifact rejection. J. Neurosci. Methods 192, 152–162. doi: 10.1016/j.jneumeth.2010.07.015
Nuwer, M. (1997). Assessment of digital EEG, quantitative EEG, and EEG brain mapping digital EEG. Neurology 49, 277–292. doi: 10.1212/WNL.49.1.277
Nuwer, M. R., Hovda, D. A., Schrader, L. M., and Vespa, P. M. (2005). Routine and quantitative EEG in mild traumatic brain injury. Clin. Neurophysiol. 116, 2001–2025. doi: 10.1016/j.clinph.2005.05.008
Oostenveld, R., Fries, P., Maris, E., and Schoffelen, J.-M. (2011). FieldTrip: open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intell. Neurosci. 2011:156869. doi: 10.1155/2011/156869
Parra, L. C., Spence, C. D., Gerson, A. D., and Sajda, P. (2005). Recipes for the linear analysis of EEG. Neuroimage 28, 326–341. doi: 10.1016/j.neuroimage.2005.05.032
Pereira, F., Mitchell, T., and Botvinick, M. (2009). Machine learning classifiers and fMRI: a tutorial overview. Neuroimage 45, S199–S209. doi: 10.1016/j.neuroimage.2008.11.007
Plis, S. M., Sarwate, A. D., Wood, D., Dieringer, C., Landis, D., Reed, C., et al. (2016). COINSTAC: a privacy enabled model and prototype for leveraging and processing decentralized brain imaging data. Front. Neurosci. 10:365. doi: 10.3389/fnins.2016.00365
Poldrack, R. A., Baker, C. I., Durnez, J., Gorgolewski, K. J., Matthews, P. M., Munafò, M. R., et al. (2017). Scanning the horizon: towards transparent and reproducible neuroimaging research. Nat. Rev. Neurosci. 18, 115–126. doi: 10.1038/nrn.2016.167
Poldrack, R. A., Barch, D. M., Mitchell, J. P., Wager, T. D., Wagner, A. D., Devlin, J. T., et al. (2013). Toward open sharing of task-based fMRI data: the OpenfMRI project. Front. Neuroinform. 7:12. doi: 10.3389/fninf.2013.00012
Prichep, L. S. (2005). Use of normative databases and statistical methods in demonstrating clinical utility of QEEG: importance and cautions. Clin. EEG Neurosci. 36, 82–87. doi: 10.1177/155005940503600207
Prichep, L. S., Ghosh Dastidar, S., Jacquin, A., Koppes, W., Miller, J., Radman, T., et al. (2014). Classification algorithms for the identification of structural injury in TBI using brain electrical activity. Comput. Biol. Med. 53, 125–133. doi: 10.1016/j.compbiomed.2014.07.011
Proudfit, G. H. (2015). The reward positivity: from basic research on reward to a biomarker for depression. Psychophysiology 52, 449–459. doi: 10.1111/psyp.12370
Robbins, T. W., Gillan, C. M., Smith, D. G., de Wit, S., and Ersche, K. D. (2012). Neurocognitive endophenotypes of impulsivity and compulsivity: towards dimensional psychiatry. Trends Cogn. Sci. 16, 81–91. doi: 10.1016/j.tics.2011.11.009
Saad, J. F., Kohn, M. R., Clarke, S., Lagopoulos, J., and Hermens, D. F. (2015). Is the theta/beta EEG Marker for ADHD inherently flawed? J. Atten. Disord. doi: 10.1177/1087054715578270 [Epub ahead of print].
Schirrmeister, R. T., Springenberg, J. T., Fiederer, L. D. J., Glasstetter, M., Eggensperger, K., Tangermann, M., et al. (2017). Deep Learning with Convolutional Neural Networks for Brain Mapping and Decoding of Movement-Related Information from the Human EEG. Available at: http://arxiv.org/abs/1703.05051
Siegel, M., Donner, T. H., and Engel, A. K. (2012). Spectral fingerprints of large-scale neuronal interactions. Nat. Rev. Neurosci. 13, 121–134. doi: 10.1038/nrn3137
Solís-Vivanco, R., Rodríguez-Violante, M., Rodríguez-Agudelo, Y., Schilmann, A., Rodríguez-Ortiz, U., and Ricardo-Garcell, J. (2015). The P3a wave: a reliable neurophysiological measure of Parkinson’s disease duration and severity. Clin. Neurophysiol. 126, 2142–2149. doi: 10.1016/j.clinph.2014.12.024
Steele, V. R., Fink, B. C., Maurer, J. M., Arbabshirani, M. R., Wilber, C. H., Jaffe, A. J., et al. (2014). Brain potentials measured during a Go/NoGo task predict completion of substance abuse treatment. Biol. Psychiatry 76, 75–83. doi: 10.1016/j.biopsych.2013.09.030
Sutton, S., Braren, M., and Zubin, J. (1965). Evoked-potential correlates of stimulus uncertainty. Science 150, 1187–1188. doi: 10.1126/science.150.3700.1187
Teeters, J. L., Harris, K. D., Millman, K. J., Olshausen, B. A., and Sommer, F. T. (2008). Data sharing for computational neuroscience. Neuroinformatics 6, 47–55. doi: 10.1007/s12021-008-9009-y
Thatcher, R. W., Biver, C. J., and North, D. M. (2003). Quantitative EEG and the Frye and Daubert standards of admissibility. Clin. EEG Neurosci. 34, 39–53. doi: 10.1177/155005940303400203
Thatcher, R. W., North, D. M., Curtin, R. T., Walker, R. A., Biver, C. J., Gomez, J. F., et al. (2001). An EEG severity index of traumatic brain injury. J. Neuropsychiatry Clin. Neurosci. 13, 77–87. doi: 10.1176/jnp.13.1.77
Thatcher, R. W., Walker, R. A., Gerson, I., and Geisler, F. H. (1989). EEG discriminant analyses of mild head trauma. Electroencephalogr. Clin. Neurophysiol. 73, 94–106. doi: 10.1016/0013-4694(89)90188-0
Wiener, M., Sommer, F. T., Ives, Z. G., Poldrack, R. A., and Litt, B. (2016). Enabling an open data ecosystem for the neurosciences. Neuron 92, 617–621. doi: 10.1016/j.neuron.2016.10.037
Keywords: EEG, open data, pattern classification, databases as topic, clinical neuroscience
Citation: Cavanagh JF, Napolitano A, Wu C and Mueen A (2017) The Patient Repository for EEG Data + Computational Tools (PRED+CT). Front. Neuroinform. 11:67. doi: 10.3389/fninf.2017.00067
Received: 03 July 2017; Accepted: 06 November 2017;
Published: 21 November 2017.
Edited by:
Pedro Antonio Valdes-Sosa, Joint China-Cuba Laboratory for Frontier Research in Translational Neurotechnology, ChinaReviewed by:
Thomas Marston Morse, Yale University, United StatesChristian O’Reilly, École Polytechnique Fédérale de Lausanne, Switzerland
Copyright © 2017 Cavanagh, Napolitano, Wu and Mueen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: James F. Cavanagh, amNhdmFuYWdoQHVubS5lZHU=