- Department of Neurology, University of Pittsburgh, Pittsburgh, PA, United States
Seizures have a profound impact on quality of life and mortality, in part because they can be challenging both to detect and forecast. Seizure detection relies upon accurately differentiating transient neurological symptoms caused by abnormal epileptiform activity from similar symptoms with different causes. Seizure forecasting aims to identify when a person has a high or low likelihood of seizure, which is related to seizure prediction. Machine learning and artificial intelligence are data-driven techniques integrated with neurodiagnostic monitoring technologies that attempt to accomplish both of those tasks. In this narrative review, we describe both the existing software and hardware approaches for seizure detection and forecasting, as well as the concepts for how to evaluate the performance of new technologies for future application in clinical practice. These technologies include long-term monitoring both with and without electroencephalography (EEG) that report very high sensitivity as well as reduced false positive detections. In addition, we describe the implications of seizure detection and forecasting upon the evaluation of novel treatments for seizures within clinical trials. Based on these existing data, long-term seizure detection and forecasting with machine learning and artificial intelligence could fundamentally change the clinical care of people with seizures, but there are multiple validation steps necessary to rigorously demonstrate their benefits and costs, relative to the current standard.
Highlights
• Seizure detection, prediction, and forecasting technologies can be evaluated based on sensitivity, false positive rate, and deficiency time.
• Performance of these technologies must be evaluated on unseen data.
• Seizure detection technologies can have high sensitivity for motor seizures but less for other seizure types.
• Seizure prediction and forecasting can substantially improve quality of life in people with epilepsy.
1 Introduction
A seizure is defined by the Oxford dictionary as a sudden attack of illness (1). The sudden nature of symptoms is a key component of the disability incurred, as measured both by reduced quality of life and increased mortality rates (2–4). For people with transient neurological events, the first question is if the event represents an epileptic seizure, functional (nonepileptic) seizure, or non-epileptic non-functional event (e.g., convulsive syncope) (5, 6). The differentiation of these transient neurological events can be challenging without simultaneous video-electroencephalographic monitoring (VEM): 30% of patients with presumed epilepsy who undergo VEM instead have functional seizures and 10% of patients who present for prolonged seizure to emergency rooms in clinical trials had functional and not epileptic seizures (7–9). VEM often requires hospitalization, tends to last for less than 10 days at a time, and is primarily available at tertiary care centers (10). Therefore, one key clinical challenge is to develop hardware and software technologies for highly accurate, reliable, and long-term detection of epileptic seizures.
In addition to differentiation of epileptic seizures from non-epileptic transient neurological events, there are substantial challenges in counting epileptic seizures in people with known epilepsy. Obtaining an accurate and reliable count of epileptic seizures is a foundational aspect of making treatment decisions for people with epilepsy because people with continued epileptic seizures may require alteration of treatment to reduce or eliminate epileptic seizures. The seizure counts provided by patients, witnesses, or care partners have been the basis for clinical decision making and rigorous evaluation of antiseizure medications (ASMs) and non-pharmaceutical treatments for decades. However, human-provided seizure counts have limited sensitivity, especially for focal unaware seizures where the lapse in formation of new memories during the seizure also means that the patient can forget the seizure itself (11, 12). In addition, retrospective seizure counts may represent the overall gestalt of the patient and care partner regarding the effectiveness of treatment. Using the principles of the placebo effect, people may underestimate seizure counts when there is confidence in the treatment (13–19). Conversely, skepticism or adverse effects may contribute to nocebo effect that overestimates seizure counts (18, 20–22). In long-term clinical practice, simulation-based studies suggest that the long-term dose and number of ASMs were not substantially impacted by reduced sensitivity and false positive rate, as long as sensitivity was at least 10% (23). In short term clinical trials that aim to determine treatment response within weeks to months, statistical power could substantially improve with seizure counts based on intracranial electroencephalographic (EEG) devices, but the impact of less invasive devices may be less (23–25). Conversely to patients with continued seizures, people who are seizure free may either continue current treatment, reduce the intensity or dose of current treatments, or withdraw treatment. While there are numerous barriers to withdrawal of ASMs in people who are seizure free, this uncertainty in counting seizures may contribute to the remarkably low rate of ASM withdrawal in people with long-term seizure freedom (26).
In addition to seizure detection, if a person with epilepsy was able to reliably predict when a seizure would occur, then they could take safety precautions (e.g., pull over an automobile) or tailor their treatment based on this risk (e.g., take a rescue medication or receive responsive neurostimulation) (27–29). The severe and unpredictable nature is one challenge in addressing the fact that people with epilepsy have a 3-to-12-fold increased risk of death for all causes, as well as Sudden Unexpected Death in Epilepsy (SUDEP) (2). Currently, the clinical care of patients with seizures focuses on reducing this risk by achieving seizure freedom.
One goal of the treatment of seizures is to allow the person with seizures to live their life as if they did not have seizures by the treatment causing both seizure freedom and no adverse effects. However, in more than one third of people with seizures, antiseizure medications fail to achieve seizure freedom (30); therefore, clinical care focuses on maximizing quality of life by reducing seizure frequency, minimizing adverse effects of medications, and addressing comorbidities or complications of seizures (4, 31, 32). In people with medication-resistant epilepsy, this quality of life could be massively improved if they could reliably predict when a seizure would not occur so they can engage in valuable and rewarding activities that otherwise would not be safe (e.g., driving and swimming) (33, 34).
It is both challenging and not useful to make seemingly definite binary (yes/no) predictions of if someone will have a seizure (33, 34). Therefore, instead of seizure prediction, the approach changed to provide individualized seizure forecasts instead of predictions (35–39). Similar to forecasting the weather, seizure forecasts aim to reliably identify these high and low risk states, while recognizing that their predictions likely are imperfect. If a person has one seizure a year in the low-risk state and one seizure a day in the high-risk state, then they could take meaningful precautions based on this forecast.
To develop these forecasts, clinicians and researchers have developed long-term monitoring hardware and software to both differentiate seizures from mimics as well as identify pre-seizure states (35, 37, 40). The hardware aims to measure signals from the person with seizures over time that are consistent, reliable, and have minimal noise so that the software can use data-driven techniques like machine learning and artificial intelligence (41, 42). Machine learning (ML) tools are trained based on historical data to maximize their performance based on a single quantitative metric (e.g., accuracy of classifying seizure [ictal], pre-seizure [pre-ictal], post-seizure [post-ictal], versus between seizure [interictal] states). Artificial intelligence (AI) tools are designed to perform a broad range of tasks, including tasks for which they have not been explicitly trained, and can do so using multiple ML tools (43). We are not aware of any AI tool for seizures that both fills that definition and is approved or cleared for clinical use by the United States Food and Drug Administration (FDA). There are numerous multiple FDA-cleared ML algorithms applied to data relevant to seizures and ML-based devices (see Sections 3–5).
While there is substantial hype about the power of ML/AI for clinical medicine (44, 45), especially seizures, there are only a limited number of tools that have rigorously demonstrated their utility and have been integrated into standard clinical practice (43). This is despite decades of effort, especially focused on seizure detection and prediction using electroencephalography (EEG) (46, 47). In order to understand the future of ML/AI for clinical practice, we must then understand the benefits and limitations of the existing approaches.
The central focus of this narrative review is highlighting the existing software and hardware for the detection and forecasting of seizures. Before discussing that existing software and hardware, we provide context by discussing key aspects of how these technologies’ performances should be evaluated by a clinician, person with seizures, or patient-advocate. Afterwards, we discuss the implications for these technologies on the design and conduct of clinical trials for treatments of seizures.
2 How to evaluate new machine learning tools for seizure detection, prediction, and forecasting
Ideally, seizures are rare events. Therefore, while there are commonalities between seizures and other applications of ML/AI, there are additional unique considerations (48).
2.1 Training, testing, and validation sets
Fundamentally, ML/AI are data-driven techniques to understand patterns in historical data that are then applied to unseen data (48). When these tools are being developed, their designers should be explicit regarding what data are used for training, testing, and validation (Figure 1). We define training data as data used to learn parameters of the model (e.g., odds ratios within logistic regression). While not necessary for all applications, testing data are used to learn hyperparameters or higher-level structures within the model that are not effectively optimized simultaneous with parameters (e.g., which ML model to use). Lastly, validation data are used to evaluate the performance of the model on “unseen” data and, thereby, validation data should never overlap with training or testing data.
Figure 1. Machine learning training, testing, and validation flowsheet. The best parameters, β, of a model that maximize a chosen quantitative metric of performance are learned based on training from training data only. After application of all pre-processing steps to the testing data without modification, the best hyperparameters, θ, of a model that maximize performance are learned based on testing data, without modification of the learned parameters, β. Lastly, the expected (E) performance is measured based on validation data after application of all pre-processing steps and applying the model with the optimized parameters, β, and hyperparameters, θ. D# reflects a numbered subset of data; argmax reflects identifying the optimal argument (arg) that maximizes (max) the performance; the vertical line, |, means “given” in mathematical notation.
A common error in the development of ML/AI is “leakage” or “peeking” where validation data leaks into the process of training or testing. For example, consider an ML tool that detects seizures based on accelerometry from a wearable device on the wrist, similar to a watch (49, 50). If developers report the performance of two ML approaches, support vector machines (51) and a neural network (52), on the “validation” data, then they can perform statistical tests to determine if one ML approach achieved superior performance. However, this “validation” data could be better described as “testing” data because they do not separately apply the better algorithm to “unseen” data. The performance on the “testing” data can be inflated as compared to truly unseen data because the developers tested two approaches and chose the better approach by “peeking” at the performance on validation data, which incurs a similar bias to multiple testing in null-hypothesis significance testing (53).
While that version of “peeking” is now less prevalent than before, a common error observed in peer review of these algorithms is “leakage” during pre-processing and feature selection (53). ML/AI tools can be described as data-hungry because their ability to make reliable predictions is non-linearly and highly dependent on the size and diversity of the training data (54). When developers have access to long-term recordings, they often can derive hundreds, thousands, or even millions of quantitative metrics, p, that could predict if an event was a seizure. Because seizures are rare, it can be more challenging to have a similarly large number of examples of seizure and not seizure, n. To train stable ML/AI models, many statistical techniques require that p be less than n and, ideally, p is much less than n. To address this, developers use their biological and technical knowledge to select features that may be reliably measured and related to seizures (e.g., placing a wearable on the arm that shakes during a motor seizure) (55), in combination with statistical techniques to find the features with the best performance (56). One simple technique can include mass univariate null-hypothesis testing, where each of the p candidate features is compared in examples of seizures as compared to not seizures (57). The subsequent ML/AI tool could erroneously base its final prediction on the combination of these top features (57). This is an error because the “validation” data “leaks” into training data: validation data is used to rank the p candidate features! This “leakage” also inflates the performance of the ML/AI tool in unpredictable ways (53); therefore, developers should be extremely explicit to clarify which data contributed to each stage of training, testing, and validation.
While we caution developers and clinicians interested in evaluating these technologies for “leakage” and “peeking,” there are good techniques to maximize the size of the training, testing, and validation datasets without acquiring three entirely independent datasets (48). In cross-validation, one dataset can be split into these components artificially according to assigned proportions (e.g., 80, 10, and 10%) or based on the number of people (e.g., leave-one-person-out). Additionally, in cyclic cross-validation, the assignment of data can cycle so that each piece of data is in the validation set once and only once (Figure 2). When there are no hyperparameters to learn and thereby no need for a “testing” set, there can be just one layer of cross-validation but when there are hyperparameters and a need for a “testing” dataset, one can perform nested cross-validation where the data are split into validation and a second dataset, then cross-validation is performed by splitting into training and testing within that second dataset to learn the hyperparameters (not pictured in Figure 2). The performance observed on the cycling validation data can then reflect the overall performance of the approach, even if the underlying ML/AI tool varies slightly from cross-validation fold to fold. This allows developers to develop high performance ML/AI tools based on large datasets, as well as validate them on data that was “unseen” by the tool. This approach of cross-validation and other similar techniques allow for development of new approaches on limited datasets, but before integration into clinical practice, these tools also should be validated with external datasets on a broad population of people (43–45).
Figure 2. Examples of common errors of (A) “leakage” and (B) “peeking” where validation data is not truly “unseen.” In (A), the validation data leaks into training by being used in feature selection to identify the features related to the outcome of interest. In (B), the best performing ML model is chosen based on performance based on the “validation” data, but there is no data left to evaluate the performance of that best ML model on “unseen” data.
The final common error that we will describe here is recognizing the internal structure in the data being from multiple people, each of whom may have had multiple seizures. Intuitively, data from the same person likely is more similar than data from different people, so if the data from one person serves within both training and validation, the models may have “seen” some aspect of the validation data. For example, using a nearest neighbor approach for seizure detection likely would identify the seizure most similar to the validation data, which probably would be from the same person! To account for this, developers can and should impose structure to the data splits of cross-validation where data are assigned to training, testing, and validation based on the person and not the individual time point. However, each person’s seizures often are stereotyped; therefore, performance can be improved substantially by understanding individual-level patterns, in addition to patterns common between multiple people. To accomplish that, training and test data for a real-time seizure detection or forecasting device could include only data from the same patient acquired before the validation data, which can be considered “pseudoprospective” validation (35, 58). This pseudoprospective validation occurs retrospectively, but by restricting to data available before the validation data, the goal is to simulate how the method would work when used prospectively (See Figures 3–7).
Figure 3. Illustrating the structure of a cyclic 10-fold cross-validation, where data is split into mutually exclusive subsets labelled D#. Model training occurs on training data only (black) and validation performance is estimated from validation data only (orange). In cyclic cross-validation, the identity of which data was validation cycles so that each subset of data is used for validation once and only once. Pooled performance across folds estimates performance of the general approach on unseen data, but each of the 10 different models likely have different learned parameters, β. When hyperparameters, θ, need to be learned, nested cross-validation can further split the black data into training and testing (pink in the Figure 1).
Figure 4. Illustration of the difference between splitting data into training and validation sets when the internal structure of the data is either maintained or modified. When the data includes 10 seizures from 10 patients, indicated by SX. Y for Seizure Y from Patient X, it would be an error to use unstructured splitting (first panel). Two appropriate methods for splitting into training and validation are illustrated. In the middle panel, we show training on data from 9 patients and validating based on the left out patient. In the right panel, we illustrate pseudo-prospective validation where the model is trained based on the first 9 seizures from each patient and validating using the last seizure from each patient.
Figure 5. Examples of the electrographic, myogenic, and electrocardiographic (ECG) signals seen for (A) a left temporal onset focal to bilateral tonic clonic seizure and (B) a functional (nonepileptic) seizure with rhythmic artifacts. A challenge of seizure detection, prediction, and forecasting technologies are to differentiate these two types of events based on recording these signals with a combination of relevant sensors. The purple arrows highlight rhythmic artifact from side-to-side movement of the head against a pillow that appear to evolve like an electrographic seizure, but they can be differentiated from an epileptic seizure based on the high amplitude field in the posterior electrodes whereas the amplitudes in the anterior electrodes are markedly lower. The red markings highlight the challenges of ECG monitoring where in (A) the tonic-clonic movements include the chest and the muscle-generated signals obscure the relatively lower amplitude signals from the heart and in (B) we highlight that the ECG was not accurately recording during the seizure, which represents deficiency time.
Figure 6. Illustration of EEG-based approaches for seizure detection, seizure prediction, and seizure forecasting that differ from conventional scalp EEG. See Table 1 and text for citations of specific technologies.
Figure 7. A GPT-4 generated illustration of a person wearing various monitoring devices that could be used for seizure detection, prediction, and forecasting. The white blanket could represent a bed pad monitoring device. The watch and bicep monitoring devices highlight where other external sensors can be placed. The headphones represent devices that can be worn around or inside the ear or head. See Table 2 and text for further descriptions.
2.2 Understanding the metrics of performance
The traditional metrics to describe the performance of ML/AI tools describe how much the ML/AI tool predictions align with a gold-standard method. Outside of seizures, these metrics can include accuracy, sensitivity, specificity, predictive values, and receiver operating curves. These metrics often focus on data where the tools aim to differentiate between two conditions with relatively similar prevalence. This often is not the case for seizures. Ideally, continuous EEG monitoring for seizures in a patient who is critically ill observes zero seizures (59). When seizures are present, there often is fewer than 2 seizures, each lasting 2 to 5 min, within a 24 h recording. Therefore, even if a patient had 10 min of seizure in 24 h, any algorithm that naively predicted the patient was always not having a seizure based on classification of 1-min periods of EEG would have an accuracy of 99.3% (wrong for only 10 of 1,440 min). However, those 10 min are critically important to the care of that patient!
To overcome this, seizure detection, prediction, and forecasting tools typically focus on precision, recall, and calibration curves (60). In ML/AI, recall is equivalent to the sensitivity: the percent of positive cases (seizures) that were identified. Precision is equivalent to the positive predictive value: the percent of predicted positive cases (seizures) that indeed were positive cases based on the gold-standard metric. Analogous to a receiver operating curve that studies the relationship between sensitivity and specificity, there are precision-recall curves that display the relationship between precision and recall based on changing risk thresholds. The exact choice of risk threshold influences the calculation of precision and recall, therefore the area under the precision-recall curve; often abbreviated AUPRC, PR-AUC, or PRC; can summarize overall performance. This AUPRC is analogous to the area under the receiver operating curve for sensitivity and specificity.
In addition to predicting seizure versus not seizure, ML/AI tools often also provide a score or risk of seizure. The user can select the hyperparameter of risk threshold: scores above this threshold are called seizures and scores below this threshold are not seizures. Higher risk thresholds will have higher recall and specificity, but lower sensitivity and precision. One method to select these thresholds are by calibration curves, which display the relationship between the predicted risk and the observed risk with gold-standard diagnosis, which is equivalent to the precision of each risk threshold. One great example of a clinically relevant calibration curve is from the 2HELPS2B algorithm that predicts risk of seizure on continuous EEG in patients who are critically ill: a score of 0 indicates <5% risk of seizure after the first hour of monitoring, whereas a score of 1 or more indicates that 12 or more hours of seizure free EEG monitoring is needed for the risk of seizure on longer monitoring to be less than 5% (59, 61, 62).
In addition to these traditional metrics, seizure detection, prediction, and forecasting tools also must recognize the quality of their data by reporting a “deficiency time” (60). When sensors are worn for long periods of time, the quality of the signal often degrades so that it falls outside the realm of the training, testing, and validation data. This concept of degradation of quality over time is familiar for people who read inpatient long term conventional EEG, where affixing electrodes with tape allows for quality data for a couple of hours but collodion often is necessary for quality data for at least 24 h. Part of the engineering challenge for other technologies is to improve or maintain the quality of the signal, as well as the durability of that quality without technologist or human intervention. When the data is of insufficient quality to use the ML/AI tool, there is a distinct clinical difference between defaulting to a prediction of not seizure, as compared to identifying poor quality data. Therefore, these ML/AI tools should report a “deficiency time,” which is the duration of time where the data was of insufficient quality for detection, prediction, or forecasting (60). Unfortunately, because the datasets used to develop these ML/AI tools often are curated to be high quality, some developers neglect to evaluate deficiency time; therefore, users should be vigilant to consider this before utilizing the tool.
Each of these metrics of performance is estimated based on the validation data, which is a finite dataset. While it can be tempting to compare algorithms’ performance across datasets based on any one metric, it also is critical to recognize the uncertainty in each metric. For example, an accuracy of 80% in a sample of 100 patients is not statistically different from an accuracy of 82% in an independent sample of 100 patients (Fisher exact test Odds Ratio 0.88, 95% confidence interval 0.43–1.78, p = 0.86). Whenever comparisons are made between performances, readers should be diligent to make sure the comparison is based on rigorous statistical testing (56).
Often, summary metrics for machine learning are not normally distributed and often their distribution is unknown (e.g., variable importance). Common techniques to overcome these limitations are permutation testing and bootstrapping. In permutation testing, the link between the predictors (input data) and gold-standard outcome is broken by shuffling the rows of the outcome vector, without replacement. When the entire process of data cleaning, feature selection, training, testing, and validation are performed on a broken or permuted dataset, then the summary output should reflect chance performance. If this process is repeated for 10,000 independent permutations of the outcome vector, then one can build an empiric probability distribution for any summary metric. The traditional null hypothesis can be tested by asking if the observed metric (e.g., variable importance) was as extreme or more extreme than 5% of observed values on the permuted dataset (p < 0.05). A rule of thumb is that for this 5% threshold to be consistent when the permutation testing is repeated, at least 10,000 independent permutations should be done. For a threshold of 1%, then 50,000 independent permutations should be done. When algorithms are computationally intensive, these rules of thumb may not be practically possible. In comparison to permutation testing that empirically estimates the distribution of chance, bootstrapping takes the opposite tactic by estimating the distribution of observed performance. Bootstrapping creates datasets by randomly selecting data points (e.g., patients or seizures) with replacement so that, on average, some data points will be selected more than once. The entire process of machine learning can occur on these bootstrapped datasets to create an empiric probability distribution for the observed performance. While we advocate for permutation testing and bootstrapping, we also acknowledge that they are not appropriate in all cases (e.g., when the dataset is too small for 10,000 separate datasets to exist).
Additionally, the best comparison between ML/AI tools are made based on benchmark datasets (63, 64), where each algorithm is validated based on the same data. If this is not possible, then developers should ensure that the datasets used to measure performance are comparable: for example, it is not appropriate to compare performance of a seizure detection tool on critical care EEG to outpatient EEG (46, 47).
3 Seizure detection with EEG and electrocorticography
There have been decades of work by exceptional researchers on how to detect seizures using quantitative EEG features, but the problem remains unsolved (65–67). In this section, we describe the existing software and hardware approaches for seizure detection using traditional EEG, as well as long-term and wearable solutions (38).
The current standard for seizure detection using traditional clinical-quality EEG in critical care is SPaRCNet, which used a large dataset of long-term critical care scalp EEGs to train a deep neural network to identify electrographic seizures and other abnormal findings (46). To train this algorithm, there were nine independent board-certified epileptologists or neurophysiologists who annotated 15 s snippets of EEG. Unfortunately, one of the challenges for seizure detection is inter-rater agreement, which was an average of 55% for each pair of annotators. Therefore, SPaRCNet used the “gold-standard” of the majority vote of these annotators, which may or may not reflect a true “gold-standard” as compared to implanted intracranial direct ECoG in the appropriate region.
In addition to SPaRCNet, it is important to mention SCORE-AI, which similarly is an ML/AI tool used to interpret outpatient scalp EEG (47). While SCORE-AI also has impressive performance in the outpatient setting, the development dataset had a very limited number of epileptic seizures; therefore, its performance to detect seizures has yet to be determined reliably. Instead, it focuses on identifying interictal abnormalities.
Additionally, the intended application for ML/AI is to work in combination with human experts. To date, these technologies have been compared to human experts, but there has not been an evaluation of the performance of human experts with and without ML/AI assistance. Therefore, prior to wide adoption of these technologies, it should be shown directly that the human assisted by the ML/AI improves upon human performance without the ML/AI. We hypothesize that humans with ML/AI will be able to read EEG studies much faster with the same quality, but that study has not been done yet.
The most widely used and FDA-approved ML/AI tool for seizure detection using long-term ECoG is applied within the Responsive Neurostimulation System (RNS) (68–72). Patients with RNS have a small number of electrodes surgically implanted into the region thought to be the seizure onset zone or another area that can modulate seizures. Traditional programming of the RNS uses a small set of quantitative EEG features to identify when a patient is having an electrographic seizure and, in real-time, provide electrical stimulation to terminate the seizure. These programming settings are initially based on collaboration between an expert epileptologist and RNS engineer to maximize the sensitivity of prediction, provide stimulation as early as possible in the seizure, and minimize false positive rates to preserve battery life. In patients whose seizures do not improve in frequency or severity after neurostimulation programmed by human experts, there is an ML-based algorithm to suggest programming settings to optimize stimulation (73). These program settings are implemented using a human-in-the-loop approach where the ML-based algorithm suggests the programming settings, but human supervision is required prior to implementing those settings for the patient. This is our first example demonstrating that ML/AI tools do not currently aim to replace human experts, but the performance of human experts can be improved by collaboration with ML/AI tools.
In addition to these two more established technologies, there are numerous technologies in development that address the challenges of the established technology (Table 1) (38, 74, 75). The placement of the electrodes for the Responsive Neurostimulation System requires localization of the seizures, which can be challenging in some cases (73). In addition to other intracranial implants, there are technologies to implant targeted EEG recording electrodes underneath the skin or other tissues (76–80). After a small incision, these devices function for at least 1 month and, based on tolerability, have been used for at least 1 year. The next step down in invasiveness is targeted non-conventional scalp EEG systems, where the electrodes can be worn inside the ear or affixed to the skin through adhesives or placed within the ear to provide a limited coverage for days and perhaps up to a week or two (81–85). These non-conventional systems aim to replace ambulatory EEG systems, which often require an EEG technologist for placement and typically are limited to up to 72 h of monitoring (86).
Table 1. Categories of EEG-based approaches for seizure detection, seizure prediction, and seizure forecasting (illustrated in Figure 6).
4 Seizure detection without EEG
One ultimate goal of seizure detection technologies is to produce automated and reliable counts and diagnoses of epileptic seizures based on long-term wearable devices (Table 2) (87–90). If a person with seizures could wear a device with high sensitivity for seizure detection and low false positive rate, then the device could alert caregivers, emergency services, and others to when the person with seizures requires assistance which, in turn, could have a direct impact upon risks of Sudden Unexpected Death in Epilepsy (SUDEP), early treatment of status epilepticus, and monitoring for treatment response (90). The challenges to seizure detection without EEG is that the device must (1) capture data relevant to the seizure, (2) reliably differentiate seizure from non-seizure, and (3) be wearable.
Table 2. Categories of non-EEG based approaches for seizure detection, seizure prediction, and seizure forecasting (illustrated in Figure 7).
Current clinical decisions are made based on quasi-objective seizure diaries where patients and their caregivers keep track of a rough seizure count. Based on this rough seizure count, clinicians attempt to judge whether seizures change in response to modifications in treatment. Analysis of the electrocorticography from the implanted Responsive Neurostimulation System indicated that treatment response could be predicted within a week of reaching the equilibrium dose of an ASM, but that requires intracranial surgery (25). If similar quality of seizure counts could be achieved with less invasive monitoring, then treatment responses could be made much more quickly and reliably.
There are many non-electrophysiological methods to monitor for seizures including accelerometry (ACM), electromyographic (EMG), cardiac monitoring, electrodermal activity (EDA) to measure sweating, and photoplethysmography (PPG) to measure blood oxygenation with light (90, 91). As discussed in Section 2, these technologies are primarily judged based on their sensitivity (what percent of seizures do they capture), false positive rate (what percent of alerts are indeed seizures), and deficiency time (what percent of the time they capture quality data) (60).
The principle behind accelerometry is that many seizures either produce movements or represent an absence of movement or tone. With objective recording with a wearable device, the detailed characteristics of these movements have been able to differentiate bilateral or generalized tonic-clonic epileptic seizures from mimics including functional seizures (otherwise known as psychogenic nonepileptic seizures) as well as other transient neurological conditions like convulsive syncope (92–94). This has been demonstrated in small to moderate size studies, but the performance of these technologies has not been demonstrated in the broad and medically complex populations typically seen in outpatient clinic or even video-EEG monitoring units. The first evaluation of these devices typically is in patients admitted in a highly controlled environment for video-EEG monitoring to directly compare the device to the gold standard. Once evaluated there, devices can then be utilized in the real world. These subsequent evaluations have shown that many behaviors like clapping, walking, running, and other repetitive motor behaviors can be challenging to differentiate from seizures. In some patients, this can produce as many as two false positive seizure detections a day, which would be intolerable in the context of having one epileptic seizure every 3 months (95, 96).
In addition to accelerometry, wearable sensors can evaluate other aspects of a person’s health (90). Smart watches are increasingly prevalent and can directly associate seizures with cardiac cycling. In fact, there is emerging work demonstrating synchronicity of both circadian and multi-day cycles of seizures and heart rate variability (97). Due to the inherent limitations of accelerometry, the addition of multimodal sensors has the great potential to improve upon both sensitivity and false positive rate. The addition of sensors also may increase the invasiveness of the technology and thereby create wearability challenges if the patient can only tolerate wearing the sensors for a restricted period. For example, bed covers are a good example of being very wearable because they only change the feel of the bed, but also miss seizures that do not occur in bed. In contrast, the Responsive Neurostimulation System (RNS) requires surgical implantation, daily to weekly data uploads, and in person clinic visits roughly every 3 months for programming. Other technologies balance between these two extremes of least invasive to most invasive. Table 2 summarizes some of the current approaches using multimodal sensors for seizure detection.
5 Seizure prediction and forecasting
While seizure detection itself is valuable because it can improve patients’, caregivers’, and the healthcare system’s responses to seizures, it would be even more powerful if we could predict when seizures would occur in the future (33, 34, 98). The practical definition of epilepsy is based upon a greater than 60% 10 year cumulative risk of unprovoked seizures (99), but that the risk of seizures is not the same on each individual hour, day, week, or year. Some of the key aspects of the disability incurred by recurrent seizures is the lack of predictability. A fear of seizure can contribute to a new onset or worsening anxiety disorder (32, 37). Similarly, the feeling of learned helplessness from being unable to control when seizures occur can contribute to the high rate of comorbid depression in people with seizures (100). If individual minutes or days with greater than 60% risk of seizures could be predicted reliably, then practical actions could be taken to improve both quality of life and safety, like taking additional medication(s), alerting a caregiver, or not driving an automobile. In days of low seizure risk (e.g., <1%), a radical suggestion would be that people with seizures could be indistinguishable from people without seizures and may not even require ASMs, similar to oligoepilepsy (101). Unfortunately, our methods for seizure prediction and forecasting are not yet good enough to achieve that vision for the future (102).
In addition to producing practically useful tools, the process of developing seizure prediction and forecasting methods improved our understanding of the continuum of states from inter-ictal, pre-ictal, ictal, to post-ictal. These states differ from the reporting guidelines for expert interpretation of EEG, which focus on differentiating inter-ictal activity from ictal-interictal continuum and electrographic or electroclinical seizure. In some of the pivotal work on seizure prediction, intracranial monitoring with NeuroVista demonstrated that in order to have a seizure, many patients first transition to a pre-ictal state where the likelihood of seizure is high, but seizure is not guaranteed (103, 104). In the inter-ictal state, the risk of transitioning directly to seizure within seconds or minutes is very low (e.g., 1/year), but the risk of subsequent seizure in the pre-ictal state is higher (e.g., 1/week). This is consistent with other observations about the propagation of the epileptic activity from the seizure onset zone to a symptomatogenic zone that produces clinical symptoms including, but not limited to, “auras” or focal aware seizures in isolation (105, 106). Many patients with focal-onset seizures have focal aware seizures that do not progress to focal unaware seizures. In contrast to focal aware seizures (auras), these pre-ictal states are not seizures. However, recognition of pre-ictal states could represent an opportunity for intervention to avoid progression to seizures (See Table 3).
Table 3. Glossary of common and important terms in the field of machine learning and artificial intelligence for seizure detection, prediction, and forecasting.
If this pre-ictal state can be identified reliably, then targeted changes to treatments can be implemented to transition back to the inter-ictal state as compared to the ictal state (33). For example, Chiang and colleagues demonstrated that the different types of neurostimulation were effective to reduce seizures when applied in a pre-ictal state as compared to an inter-ictal state (107). Even when seizures do not occur in someone who is seizure free, identification of these transitions to a pre-ictal state could delineate the risk of seizure with medication withdrawal. Recruiting for randomized trials for medication withdrawal can be challenging because if someone is seizure and side effect free on antiseizure medication(s), then there may not be motivation to stop the antiseizure medication(s). Lowering or stopping antiseizure medications in this context can reduce healthcare costs and adverse effect burden, but there’s always a risk of breakthrough seizures when coming off medications and even one breakthrough seizure can have big implications (e.g., driving restriction, injuries from seizures, and SUDEP). In addition, it can be challenging to train a machine learning algorithm to identify seizures in patients who are seizure free because there may be no examples of an individual’s seizures captured with the device for training (see Section 2 for discussion of internal structures in data). While training based on others’ seizures can somewhat translate to individual patients, these seizure prediction and forecasting methods often require personalized modifications (26). Especially because there is no electrographic definition of a pre-ictal state, it is not yet possible to reliably identify this preceding state without observing the seizure that it precedes.
These concepts of state transitions have led the field to transition from seizure prediction to seizure forecasting (34). In seizure prediction, technologies aim to make clear yes or no predictions that a seizure will or will not occur. In seizure forecasting, the purpose is to provide a risk-gauge for seizures that likely differs from these clear yes and no insights. Like weather forecasting, these seizure forecasts are most helpful when the risk substantially varies with time so that different decisions can be made based on this estimated risk, but also recognizing that the predictions are imperfect.
One of the insights from seizure forecasting is that continuous technological monitoring may not be necessary to provide reasonable seizure forecasts (98). Part of the phenomenology of Juvenile Myoclonic Epilepsy is the circadian cycle where myoclonic jerks and seizures are more common in the early morning. Those circadian cycles also are present in seizures observed in continuous video-EEG monitoring units and critical care (40, 103, 104, 108, 109). However, these circadian cycles may not be as prominent as suggested by the data, because there may be a confound of circadian under-reporting of seizures (e.g., nocturnal focal aware seizures are missed because patients are sleeping, causing the data to suggest that focal aware seizures only occur during the day) (110).
In addition to circadian rhythms, the multi-day cycles in seizures have been well recognized (103, 108, 111). Catamenial or menstrual cycle related seizure patterns have long been recognized based on patient and caregiver observations, with subsequent work demonstrating sub-patterns of catamenial seizures based on the relative sensitivity to the pro-convulsant effect of estrogen and the anti-convulsant effect of progesterone (112, 113). Recognition of these patterns can prompt cyclic prescription of antiseizure medications, where a higher dose is taken 3 days before, during, and 3 days after the “high risk” period (113).
The addition of long-term wearable technologies has enhanced our recognition of these seizure cycles being present in as high as 30% of people with medication resistant epilepsy (111, 114). When seizure diaries were paired with long-term cardiac monitoring, Cook and colleagues recognized that the monthly cycles of seizures occur in men with a similar prevalence as catamenial epilepsy (97). They also observed synchrony of the heart rate variability in concurrent cardiac monitoring with a watch, as compared to seizure cycles, which prompted hypotheses about the mechanisms of SUDEP (2). In addition, there commonly are empiric cycles with differing lengths, including weekly seizures, fortnightly seizures, and even multi-month patterns (111, 114). When these patterns are identified in a person, then additional clinical questions arise regarding potential cyclic non-adherence to medications (e.g., seizures on Mondays after missing weekend antiseizure medication doses) or other associations with external or internal factors (e.g., sleep deprivation, head injuries, hormonal changes). Other multi-day fluctuations may be based on the body’s intrinsic and adaptive resilience to seizures through balancing excitation and inhibition within epileptic networks.
An important concept to identifying these seizure patterns is the Nyquist frequency, which some providers may be familiar with based on interpreting EEG. The Nyquist frequency states that to observe a signal with a certain frequency, then the sampling rate must be at least twice as fast or twice as long as the target frequency. In EEG, this means that frequencies faster than 125 Hz cannot be observed if the EEG data has a sampling frequency of 250 Hz. For seizure cycles, to measure a monthly seizure cycle, there must be at least 2 months of data. Based on this concept, devices that use monthly cycles for seizure prediction and forecasting must be worn for at least 2 months before there is sufficient data from the individual patient to measure that cycling. This represents a technological minimum and the reliability of measuring these cycles can improve if longer monitoring is performed. To highlight that the Nyquist frequency is the technological minimum, the International League Against Epilepsy uses other statistical principles to define a response to treatment by an at least tripling of the between-seizure interval, which can be called the “Rule of Three.” (99).
The above concepts of seizure prediction and forecasting primarily are within the realm of research and have not been demonstrated as directly clinically applicable yet (102). When thinking about the future of machine learning and artificial intelligence, we recognize that the current existing tools are the worst that we will see in the future (43). The goal of researchers, engineers, and clinicians is to use our constantly improving capabilities for data processing and understanding to improve upon these standards. While healthcare is commonly very slow to migrate, we recognize the great capacity for machine learning and artificial intelligence to make massive disruptive changes. For example, the public release of large language models like ChatGPT have forever revolutionized the practice of essay writing in higher education (115, 116). Even though these disruptive technologies do not exist for seizure detection, prediction, and forecasting yet, we must begin to understand both their potential benefits and limits so that when, not if, these technologies exist, then we are able to utilize them responsibly (43).
6 Implications for clinical trials
One initial use for these technologies for the detection, prediction, and forecasting of seizures is in clinical trials of treatments for epilepsy. Clinical trials are the foundation upon which we determine the efficacy of novel treatments, but there are increasing challenges to the design and conduct of trials (14). These challenges are highlighted by the reducing number of participants per site recruited, which requires a compensatory increase in the number of sites to meet sample size goals (14).
However, this increase in sites also has been associated with a progressive increase in the placebo response rate, which likely harms the statistical power of trials by shrinking the difference between the elevated placebo and the efficacious treatment, which may be capped by a ceiling effect (14, 16). That creates a vicious cycle where higher placebo response rate reduces statistical power, which increases the number of participants and sites which, in turn, further increases the placebo response rate.
The insights that we have gained from seizure detection, prediction, and forecasting work have both highlighted those problems as well as offered potential solutions. The challenge in recruiting participants in a site is that more than 50% of patients with medication resistant epilepsy seen at a tertiary care center for epilepsy with a long history of participating in clinical trials did not meet the minimum seizure frequency requirement (117). To be able to observe the efficacy of treatment within a reasonable time frame of a 12 or 20 week maintenance period, focal-and generalized-onset seizure trials often require at least 4 seizures per month and 3 seizures per 2 months, respectively. Lowering those minimum seizure frequency eligibility requirements may necessitate longer maintenance periods, but longer maintenance periods further increase cost and hinder participant recruitment because placebo exposure in blinded trials was associated with a 5.8-fold increased risk of SUDEP (118).
One of the challenges incurred by these minimum seizure frequency requirements called regression to the mean can be understood using the language of seizure forecasting (119). In regression to the mean, a participant is motivated to enroll in a placebo-controlled trial of a new antiseizure medication due to a progressive recent increase in seizure frequency from 1 per month to 4 per month for 3 months (119). Now that they are eligible to participate, they start keeping a prospective seizure diary during the 4-to-8-week pre-randomization baseline. If this recent increase in seizure frequency was due to inherent changes in their seizure cycles causing an increase in their rolling-average seizure forecast of 4/month, then they may spontaneously transition back to that low seizure frequency state (1/month) during the baseline and be a “baseline failure.” However, if this spontaneous transition occurs after the pre-randomization baseline and instead during the blinded randomization phase, then their 75% reduction in seizure frequency from 4/month to 1/month may be unrelated to the efficacy of their blinded treatment. While trials do not report the reason for “baseline failures,” the increase in placebo responder rate has been associated with an increasing rate of baseline failures (14).
In addition to giving us the language to describe those changes, seizure forecasting can provide solutions. The seizure cycling literature demonstrated that, by and large, most seizure cycles are less than 1 month long. Therefore, the minimum duration to evaluate the efficacy of a treatment likely is longer than 1 month. Trial design can use this knowledge to create adaptive designs, where participants engage in blinded treatment until there is sufficient data to conclude that they have either responded or not responded to the blinded treatment, irrespective of the length of observation. One proposal to accomplish this is through the Time to Prerandomization monthly Seizure Count (T-PSC) design, where participants engage in blinded treatment until their blinded post-randomization seizure count exceeds their average monthly seizure count measured during baseline (13, 120). This design allows for quick exclusion of participants who worsen on blinded treatment (either due to nocebo or worsening of seizures on novel treatment), who likely have an increased risk of SUDEP and other serious adverse events (118). [The nocebo effect is defined as a negative effect of placebo treatment, including either worsening of seizures or adverse effects (20)]. Simultaneously, this design also either maintains the traditional 12-to-20-week endpoint for initial responders. This traditional 12-to-20-week endpoint remains necessary to differentiate novel treatments from short-acting benzodiazepines, which transiently lower seizure frequency but do not have a sustained effect. Re-analyses of numerous historical trials has demonstrated that the efficacy conclusions of the trials truncated at T-PSC were almost identical to the conclusions with full-length trial data (13).
The other challenge of clinical trials is producing reliable and accurate counts of epileptic seizures (11). Foundationally, the approval of novel treatments is based upon demonstrating both the magnitude of the difference in efficacy between placebo and active treatment. While patient-and care partner-reported diaries are the current standard for clinical trials, there is substantial data suggesting that the sensitivity of these reported seizure detections is less than 50%, where some seizure types may have even lower sensitivity (e.g., focal aware seizures, nonmotor focal unaware seizures, absence seizures) (24). There also may be circadian patterns in reduced sensitivity that may create artificial circadian patterns in seizure diary data (e.g., reduced sensitivity during sleep) (110). Theoretically, this reduced sensitivity increases the variability of seizure count by adding a source of noise, as well as lowering the apparent seizure frequency in each phase of the trial, which can impact the certainty in measuring each seizure frequency as well as eligibility (e.g., insufficient baseline seizures to progress to the blinded phase).
In addition to increasing the variability of the difference estimate within trials, imperfect human-provided seizure counts can incur bias (110). While bilateral or generalized tonic-clonic seizures likely have low false positive rate of reporting, data from ambulatory and inpatient EEG monitoring suggest that patients and care partners also may have an unclear false positive rate due to inaccurate differentiation of other transient neurological symptoms from their seizures (121). Especially for scalp EEG negative focal aware seizures, it can be very challenging to evaluate seizure from non-seizure events. Theoretically, false positive seizure counts cause under-estimation of difference between placebo and active treatment because true seizures may benefit from treatment, but false positive counts would not.
These limitations in human-provided seizure diaries clearly could be assisted by objective devices that improve upon both the sensitivity and false positive rates. Long-term wearable or implanted devices could provide more accurate time stamps on seizures and provide further characterization of many details regarding seizures (e.g., duration of the motor phase) that often is too burdensome for humans to reliably record. Combined with the development of seizure prediction and forecasting methods, these granular data can incorporate the known confounding factors of seizure cycling and clustering into improved quantitative metrics measuring improvements in seizures.
To achieve these future benefits, many devices have been shown to improve the sensitivity of seizure detection but, unfortunately, these devices also incur a false positive rate that may be higher than humans (90). Additionally, some of the long term intracranial, subcutaneous, or other EEG-based devices may reliably measure electrographic seizures, but there is an incomplete link between electrographic and electroclinical seizures (74). Improvement in electrographic seizures likely does not guarantee an improvement in electroclinical seizures, the latter of which likely have a greater impact upon quality of life. These devices also often have demonstrated this performance in selected patients who are interested in enrolling in studies of the devices, as compared to a broader population who would enroll in a trial for a novel treatment.
Each of these concerns can be addressed through further study and validation of existing devices, as well as improving upon existing technology. Lastly, clinical trials are both very expensive and high risk for sponsors who often seek to mitigate that risk through following conventional trial protocols. While integrating devices into trials could have some benefits, providing innovative devices to trial participants would increase cost, as well as risks. When (not if) a device for seizure monitoring is used in a trial, there is a nontrivial risk that the human-provided seizure outcome may differ from the device-based diary. (To our knowledge, such a trial has not been completed and published yet but, to our knowledge, there indeed is a trial underway that combines both human diaries and long-term EEG based seizure detections). To describe this risk, consider the options for when the human and device-based diaries disagree. If the human-provided diary suggests benefit but the device does not, then clinicians, sponsors, and regulators may conclude that the treatment benefitted clinically meaningful seizures because humans counted them. However, the trial would conclude that the treatment does not have an impact on the device-based seizure counts, which may be felt to not be clinically meaningful. Conversely, if only the device demonstrated treatment benefit, then the conclusion would be that this likely would be a statistically significant, but not clinically significant, improvement. This risk of discrepancy could be mitigated by using a human-in-the-loop system where the device’s seizure predictions are provided to the patient and care partner for supervision, as well as subsequent review by study staff or treating clinicians (122). However, these human-in-the-loop studies have not been done, to date, within epilepsy or seizures.
As researchers with expertise both in machine learning and clinical trials, we believe that clinical trials will eventually rely upon seizure counts assisted by long-term seizure detection and monitoring devices that use innovative machine learning techniques to improve upon the current standard of human-provided seizure diaries. Each of the concerns that we have raised and others that we have heard are addressable through further development of both the hardware and software for seizure detection, prediction, and forecasting.
7 Conclusion
ML and AI will revolutionize the clinical care of people with seizures, but the field is nascent. In the future, both software and hardware will improve the reliability of the original diagnosis of epilepsy compared to mimics, monitoring of seizure recurrence with treatment, evaluating novel treatments within clinical trials, and providing a real-time and actionable warning system that people with seizures can use to improve safety and quality of life. While these improvements are promising, the performance of the existing technology has not been demonstrated to be high enough to warrant a change in current clinical practices. We are confident that in the near future ML and AI will not replace clinicians, but clinicians assisted by ML and AI will replace clinicians not utilizing ML and AI.
Author contributions
WK: Conceptualization, Funding acquisition, Writing – original draft, Writing – review & editing. KM: Conceptualization, Writing – review & editing. GF: Conceptualization, Visualization, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the American Epilepsy Society, the Epilepsy Foundation, the American Academy of Neurology, the American Brain Foundation, and the Epilepsy Study Consortium. None of these funders reviewed this manuscript or commissioned this work specifically.
Conflict of interest
WK has received compensation for review articles for Medlink Neurology and consulting for SK Life Sciences, Biohaven Pharmaceuticals, Cerebral Therapeutics, Jazz Pharmaceuticals, EpiTel, UCB Pharmaceuticals, Azurity Pharmaceuticals, and the Epilepsy Study Consortium; and has collaborative or data use agreements with Eisai, Janssen, Radius Health, and Neureka.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Kerr, WT, and Stern, JM. We need a functioning name for PNES: consider dissociative seizures. Epilepsy Behav. (2020) 105:107002. doi: 10.1016/j.yebeh.2020.107002
2. Giussani, G, Falcicchio, G, la Neve, A, Costagliola, G, Striano, P, Scarabello, A, et al. Sudden unexpected death in epilepsy. A critical view of the literature. Epilepsia Open. (2023) 8:728–57. doi: 10.1002/epi4.12722
3. LaFrance, WC, Ranieri, R, Bamps, Y, Stoll, S, Sahoo, SS, Welter, E, et al. Comparison of common data elements from the managing epilepsy well (MEW) network integrated database and a well-characterized sample with nonepileptic seizures. Epilepsy Behav. (2015) 45:136–41. doi: 10.1016/j.yebeh.2015.02.021
4. Johnson, EK, Jones, JE, Seidenberg, M, and Hermann, BP. The relative impact of anxiety, depression, and clinical seizure features on health-related quality of life in epilepsy. Epilepsia. (2004) 45:544–50. doi: 10.1111/j.0013-9580.2004.47003.x
5. Chen, M, Jamnadas-Khoda, J, Broadhurst, M, Wall, M, Grünewald, R, Howell, SJL, et al. Value of witness observations in the differential diagnosis of transient loss of consciousness. Neurology. (2019) 92:7017. doi: 10.1212/WNL.0000000000007017
6. Wardrope, A, Jamnadas-Khoda, J, Broadhurst, M, Grünewald, RA, Heaton, TJ, Howell, SJ, et al. Machine learning as a diagnostic decision aid for patients with transient loss of consciousness. Neurol Clin Pract. (2019) 10:96. doi: 10.1212/CPJ.0000000000000726
7. Jungilligens, J, Michaelis, R, and Popkirov, S. Misdiagnosis of prolonged psychogenic non-epileptic seizures as status epilepticus: epidemiology and associated risks. J Neurol Neurosurg Psychiatry. (2021) 92:1341–5. doi: 10.1136/jnnp-2021-326443
8. Kerr, WT, Janio, EA, Chau, AM, Braesch, CT, Le, JM, Hori, JM, et al. Objective score from initial interview identifies patients with probable dissociative seizures. Epilepsy Behav. (2020) 113:107525. doi: 10.1016/j.yebeh.2020.107525
9. Kerr, WT, Anderson, A, Lau, EP, Cho, AY, Xia, H, Bramen, J, et al. Automated diagnosis of epilepsy using EEG power spectrum. Epilepsia. (2012) 53:e189–92. doi: 10.1111/j.1528-1167.2012.03653.x
10. Adenan, MH, Khalil, M, Loh, KS, Kelly, L, Shukralla, A, Klaus, S, et al. A retrospective study of the correlation between duration of monitoring in the epilepsy monitoring unit and diagnostic yield. Epilepsy Behav. (2022) 136:108919. doi: 10.1016/j.yebeh.2022.108919
11. Karoly, PJ, Romero, J, Cook, MJ, Freestone, DR, and Goldenholz, DM. When can we trust responders? Serious concerns when using 50% response rate to assess clinical trials. Epilepsia. (2019) 60:e99–e103. doi: 10.1111/epi.16321
12. LaGrant, B, Goldenholz, DM, Braun, M, Moss, RE, and Grinspan, ZM. Patterns of recording epileptic spasms in an electronic seizure diary compared with video-EEG and historical cohorts. Pediatr Neurol. (2021) 122:27–34. doi: 10.1016/j.pediatrneurol.2021.04.008
13. Kerr, WT, Auvin, S, van der Geyten, S, Kenney, C, Novak, G, Fountain, NB, et al. Time-to-event clinical trial designs: existing evidence and remaining concerns. Epilepsia. (2023) 64:1699–708. doi: 10.1111/epi.17621
14. Kerr, WT, Reddy, AS, Seo, SH, Kok, N, Stacey, WC, Stern, JM, et al. Increasing challenges to trial recruitment and conduct over time. Epilepsia. (2023) 64:2625–34. doi: 10.1111/epi.17716
15. Romero, J, Larimer, P, Chang, B, Goldenholz, SR, and Goldenholz, DM. Natural variability in seizure frequency: implications for trials and placebo. Epilepsy Res. (2020) 162:106306. doi: 10.1016/j.eplepsyres.2020.106306
16. Goldenholz, DM, and Goldenholz, SR. Placebo in epilepsy. Int Rev Neurobiol. (2020) 153:231–66. doi: 10.1016/bs.irn.2020.03.033
17. Fureman, BE, Friedman, D, Baulac, M, Glauser, T, Moreno, J, Dixon-Salazar, T, et al. Reducing placebo exposure in trials: considerations from the research roundtable in epilepsy. Neurology. (2017) 89:1507–15. doi: 10.1212/WNL.0000000000004535
18. Chamsi-Pasha, M, Albar, MA, and Chamsi-Pasha, H. Minimizing nocebo effect: pragmatic approach. Avicenna J Med. (2017) 7:139–43. doi: 10.4103/ajm.AJM_59_17
19. Goldenholz, DM, and Goldenholz, SR. Response to placebo in clinical epilepsy trials—old ideas and new insights. Epilepsy Res. (2016) 122:15–25. doi: 10.1016/j.eplepsyres.2016.02.002
20. Zis, P, Shafiq, F, and Mitsikostas, DD. Nocebo effect in refractory partial epilepsy during pre-surgical monitoring: systematic review and meta-analysis of placebo-controlled clinical trials. Seizure. (2017) 45:95–9. doi: 10.1016/j.seizure.2016.12.003
21. Zaccara, G, Giovannelli, F, Giorgi, FS, Franco, V, and Gasparini, S. Analysis of nocebo effects of antiepileptic drugs across different conditions. J Neurol. (2016) 263:1274–9. doi: 10.1007/s00415-015-8018-7
22. Zaccara, G, Giovannelli, F, and Schmidt, D. Placebo and nocebo responses in drug trials of epilepsy. Epilepsy Behav. (2015) 43:128–34. doi: 10.1016/j.yebeh.2014.12.004
23. Goldenholz, DM, Karoly, PJ, Viana, PF, Nurse, E, Loddenkemper, T, Schulze-Bonhage, A, et al. Minimum clinical utility standards for wearable seizure detectors: a simulation study. Epilepsia. (2024) 65:1017–28. doi: 10.1111/epi.17917
24. Goldenholz, DM, Tharayil, JJ, Kuzniecky, R, Karoly, P, Theodore, WH, and Cook, MJ. Simulating clinical trials with and without intracranial EEG data. Epilepsia Open. (2017) 2:156–61. doi: 10.1002/epi4.12038
25. Quraishi, IH, Mercier, MR, Skarpaas, TL, and Hirsch, LJ. Early detection rate changes from a brain-Responsive Neurostimulation System predict efficacy of newly added antiseizure drugs. Epilepsia. (2020) 61:138–48. doi: 10.1111/epi.16412
26. Terman, SW, van Griethuysen, R, Rheaume, CE, Slinger, G, Haque, AS, Smith, SN, et al. Antiseizure medication withdrawal risk estimation and recommendations: a survey of American Academy of Neurology and EpiCARE members. Epilepsia Open. (2023) 8:386–98. doi: 10.1002/epi4.12696
27. Frei, MG, Zaveri, HP, Arthurs, S, Bergey, GK, Jouny, CC, Lehnertz, K, et al. Controversies in epilepsy: debates held during the fourth international workshop on seizure prediction. Epilepsy Behav. (2010) 19:4–16. doi: 10.1016/j.yebeh.2010.06.009
28. Schulze-Bonhage, A, Sales, F, Wagner, K, Teotonio, R, Carius, A, Schelle, A, et al. Views of patients with epilepsy on seizure prediction devices. Epilepsy Behav. (2010) 18:388–96. doi: 10.1016/j.yebeh.2010.05.008
29. Mormann, F, Kreuz, T, Rieke, C, Andrzejak, RG, Kraskov, A, David, P, et al. On the predictability of epileptic seizures. Clin Neurophysiol. (2005) 116:569–87. doi: 10.1016/j.clinph.2004.08.025
30. Kwan, P, Arzimanoglou, A, Berg, AT, Brodie, MJ, Allen Hauser, W, Mathern, G, et al. Definition of drug resistant epilepsy: consensus proposal by the ad hoc task force of the ILAE commission on therapeutic strategies. Epilepsia. (2010) 51:1069–77. doi: 10.1111/j.1528-1167.2009.02397.x
31. Terman, SW, Kirkpatrick, L, Kerr, WT, Akiyama, LF, Baajour, W, Atilgan, D, et al. Challenges and directions in epilepsy diagnostics and therapeutics: proceedings of the 17th epilepsy therapies and diagnostics development conference. Epilepsia. (2023) 65:846–60. doi: 10.1111/epi.17875
32. Kanner, AM. Management of psychiatric and neurological comorbidities in epilepsy. Nat Rev Neurol. (2016) 12:106–16. doi: 10.1038/nrneurol.2015.243
33. Moss, A, Moss, E, Moss, R, Moss, L, Chiang, S, and Crino, P. A patient perspective on seizure detection and forecasting. Front Neurol. (2022) 13:779551. doi: 10.3389/fneur.2022.779551
34. Grzeskowiak, CL, and Dumanis, SB. Seizure forecasting: patient and caregiver perspectives. Front Neurol. (2021) 12:717428. doi: 10.3389/fneur.2021.717428
35. Karoly, PJ, Cook, MJ, Maturana, M, Nurse, ES, Payne, D, Brinkmann, BH, et al. Forecasting cycles of seizure likelihood. Epilepsia. (2020) 61:776–86. doi: 10.1111/epi.16485
36. Gleichgerrcht, E, Dumitru, M, Hartmann, DA, Munsell, BC, Kuzniecky, R, Bonilha, L, et al. Seizure forecasting using machine learning models trained by seizure diaries. Physiol Meas. (2022) 43:124003. doi: 10.1088/1361-6579/aca6ca
37. Stirling, RE, Cook, MJ, Grayden, DB, and Karoly, PJ. Seizure forecasting and cyclic control of seizures. Epilepsia. (2021) 62:S2–S14. doi: 10.1111/epi.16541
38. Fu, A, and Lado, FA. Seizure detection, prediction, and forecasting. J Clin Neurophysiol. (2024) 41:207–13. doi: 10.1097/WNP.0000000000001045
39. Brinkmann, BH, Karoly, PJ, Nurse, ES, Dumanis, SB, Nasseri, M, Viana, PF, et al. Seizure diaries and forecasting with wearables: epilepsy monitoring outside the clinic. Front Neurol. (2021) 12:690404. doi: 10.3389/fneur.2021.690404
40. Karoly, PJ, Ung, H, Grayden, DB, Kuhlmann, L, Leyde, K, Cook, MJ, et al. The circadian profile of epilepsy improves seizure forecasting. Brain. (2017) 140:2169–82. doi: 10.1093/brain/awx173
41. Abbasi, B, and Goldenholz, DM. Machine learning applications in epilepsy. Epilepsia. (2019) 60:2037–47. doi: 10.1111/epi.16333
42. Kaur, T, Diwakar, A, Kirandeep,, Mirpuri, P, Tripathi, M, Chandra, PS, et al. Artificial Intelligence in Epilepsy. Neurol India. (2021) 69:560–6. doi: 10.4103/0028-3886.317233
43. Kerr, WT, and McFarlane, KN. Machine learning and artificial intelligence applications to epilepsy: a review for the practicing Epileptologist. Curr Neurol Neurosci Rep. (2023) 23:869–79. doi: 10.1007/s11910-023-01318-7
44. Kulkarni, PA, and Singh, H. Artificial intelligence in clinical diagnosis: opportunities, challenges, and hype. JAMA. (2023) 330:317–8. doi: 10.1001/jama.2023.11440
45. Xu, Z, Biswas, B, Li, L, and Amzal, B. AI/ML in precision medicine: a look beyond the hype. Ther Innov Regul Sci. (2023) 57:957–62. doi: 10.1007/s43441-023-00541-1
46. Jing, J, Ge, W, Hong, S, Fernandes, MB, Lin, Z, Yang, C, et al. Development of expert-level classification of seizures and rhythmic and periodic patterns during EEG interpretation. Neurology. (2023) 100:e1750–62. doi: 10.1212/WNL.0000000000207127
47. Tveit, J, Aurlien, H, Plis, S, Calhoun, VD, Tatum, WO, Schomer, DL, et al. Automated interpretation of clinical electroencephalograms using artificial intelligence. JAMA Neurol. (2023) 80:805–12. doi: 10.1001/jamaneurol.2023.1645
48. Poldrack, RA, Huckins, G, and Varoquaux, G. Establishment of best practices for evidence for prediction: a review. JAMA Psychiatry. (2020) 77:534–40. doi: 10.1001/jamapsychiatry.2019.3671
49. Onorati, F, Regalia, G, Caborni, C, LaFrance, WC, Blum, AS, Bidwell, J, et al. Prospective study of a multimodal convulsive seizure detection wearable system on pediatric and adult patients in the epilepsy monitoring unit. Front Neurol. (2021) 12:724904. doi: 10.3389/fneur.2021.724904
50. Onorati, F, Regalia, G, Caborni, C, Migliorini, M, Bender, D, Poh, MZ, et al. Multicenter clinical assessment of improved wearable multimodal convulsive seizure detectors. Epilepsia. (2017) 58:1870–9. doi: 10.1111/epi.13899
51. Cortes, C, and Vapnik, V. Support-vector networks. Mach Learn. (1995) 20:273–97. doi: 10.1007/BF00994018
52. LeCun, Y, Bengio, Y, and Hinton, G. Deep learning. Nature. (2015) 521:436–44. doi: 10.1038/nature14539
53. Pulini, AA, Kerr, WT, Loo, SK, and Lenartowicz, A. Classification accuracy of neuroimaging biomarkers in attention-deficit/hyperactivity disorder: effects of sample size and circular analysis. Biol Psychiatry Cogn Neurosci Neuroimaging. (2019) 4:108–20. doi: 10.1016/j.bpsc.2018.06.003
54. Varoquaux, G. Cross-validation failure: small sample sizes lead to large error bars. NeuroImage. (2018) 180:68–77. doi: 10.1016/j.neuroimage.2017.06.061
55. Li, W, Wang, G, Lei, X, Sheng, D, Yu, T, and Wang, G. Seizure detection based on wearable devices: a review of device, mechanism, and algorithm. Acta Neurol Scand. (2022) 146:723–31. doi: 10.1111/ane.13716
56. Kerr, WT, Douglas, PK, Anderson, A, and Cohen, MS. The utility of data-driven feature selection: re: Chu et al. 2012. NeuroImage. (2014) 84:1107–10. doi: 10.1016/j.neuroimage.2013.07.050
57. Bennett, CM, Baird, AA, Miller, MB, and Wolford, GL. Neural correlates of interspecies perspective taking in the post-mortem Atlantic Salmon: An argument for multiple comparisons correction, in human brain mapping. Hum. Brain Mapp. (2009).
58. Rao, SR, Slater, JD, and Kalamangalam, GP. A simple clinical score for prediction of nonepileptic seizures. Epilepsy Behav. (2017) 77:50–2. doi: 10.1016/j.yebeh.2017.09.005
59. Struck, AF, Ustun, B, Ruiz, AR, Lee, JW, LaRoche, SM, Hirsch, LJ, et al. Association of an electroencephalography-based risk score with Seizure probability in hospitalized patients. JAMA Neurol. (2017) 74:1419–24. doi: 10.1001/jamaneurol.2017.2459
60. Beniczky, S, Wiebe, S, Jeppesen, J, Tatum, WO, Brazdil, M, Wang, Y, et al. Automated seizure detection using wearable devices: a clinical practice guideline of the international league against epilepsy and the International Federation of Clinical Neurophysiology. Epilepsia. (2021) 62:632–46. doi: 10.1111/epi.16818
61. Moffet, EW, Subramaniam, T, Hirsch, LJ, Gilmore, EJ, Lee, JW, Rodriguez-Ruiz, AA, et al. Validation of the 2HELPS2B seizure risk score in acute brain injury patients. Neurocrit Care. (2020) 33:701–7. doi: 10.1007/s12028-020-00939-x
62. Struck, AF, Tabaeizadeh, M, Schmitt, SE, Ruiz, AR, Swisher, CB, Subramaniam, T, et al. Assessment of the validity of the 2HELPS2B score for inpatient seizure risk prediction. JAMA Neurol. (2020) 77:500–7. doi: 10.1001/jamaneurol.2019.4656
63. Tang, J, el Atrache, R, Yu, S, Asif, U, Jackson, M, Roy, S, et al. Seizure detection using wearable sensors and machine learning: setting a benchmark. Epilepsia. (2021) 62:1807–19. doi: 10.1111/epi.16967
64. Dohmatob, ED, Gramfort, A, Thirion, B, and Varoquaux, G. Benchmarking solvers for TV-ℓ1 least-squares and logistic regression in brain imaging, Pattern Recognition in Neuroimaging. (2014): Tubingen. 1–4
65. Henriksen, J, Remvig, LS, Madsen, RE, Conradsen, I, Kjaer, TW, Thomsen, CE, et al. Automated seizure detection: going from sEEG to iEEG, 32nd Annual International Conference of the IEEE EMBS. (2010): Buenos Aires, Argentina. p. 2431–2434
66. Tzallas, AT, Tsipouras, MG, and Fotiadis, DI. Epileptic seizure detection in EEGs using time-frequency analysis. IEEE Trans Inf Technol Biomed. (2009) 13:703–10. doi: 10.1109/TITB.2009.2017939
67. Päivinen, N, Lammi, S, Pitkänen, A, Nissinen, J, Penttonen, M, and Grönfors, T. Epileptic seizure detection: a nonlinear viewpoint. Comput Methods Prog Biomed. (2005) 79:151–9. doi: 10.1016/j.cmpb.2005.04.006
68. Abouelleil, M, Deshpande, N, and Ali, R. Emerging trends in neuromodulation for treatment of drug-resistant epilepsy. Front Pain Res. (2022) 3:839463. doi: 10.3389/fpain.2022.839463
69. Starnes, K, Miller, K, Wong-Kisiel, L, and Lundstrom, BN. A review of Neurostimulation for epilepsy in pediatrics. Brain Sci. (2019) 9:283. doi: 10.3390/brainsci9100283
70. Kokkinos, V, Sisterson, ND, Wozny, TA, and Richardson, RM. Association of Closed-Loop Brain Stimulation Neurophysiological Features with Seizure Control among Patients with Focal Epilepsy. JAMA Neurol. (2019) 76:800–8. doi: 10.1001/jamaneurol.2019.0658
71. Bergey, GK, Morrell, MJ, Mizrahi, EM, Goldman, A, King-Stephens, D, Nair, D, et al. Long-term treatment with responsive brain stimulation in adults with refractory partial seizures. Neurology. (2015) 84:810–7. doi: 10.1212/WNL.0000000000001280
72. Sohal, VS, and Sun, FT. Responsive neurostimulation suppresses synchronized cortical rhythms in patients with epilepsy. Neurosurg Clin N Am. (2011) 22:481–8, vi. doi: 10.1016/j.nec.2011.07.007
73. Boddeti, U, McAfee, D, Khan, A, Bachani, M, and Ksendzovsky, A. Responsive Neurostimulation for seizure control: current status and future directions. Biomedicines. (2022) 10:112677. doi: 10.3390/biomedicines10112677
74. Haneef, Z, Yang, K, Sheth, SA, Aloor, FZ, Aazhang, B, Krishnan, V, et al. Sub-scalp electroencephalography: a next-generation technique to study human neurophysiology. Clin Neurophysiol. (2022) 141:77–87. doi: 10.1016/j.clinph.2022.07.003
75. Ren, Z, Han, X, and Wang, B. The performance evaluation of the state-of-the-art EEG-based seizure prediction models. Front Neurol. (2022) 13:1016224. doi: 10.3389/fneur.2022.1016224
76. Duun-Henriksen, J, Baud, M, Richardson, MP, Cook, M, Kouvas, G, Heasman, JM, et al. A new era in electroencephalographic monitoring? Subscalp devices for ultra-long-term recordings. Epilepsia. (2020) 61:1805–17. doi: 10.1111/epi.16630
77. Pathmanathan, J, Kjaer, TW, Cole, AJ, Delanty, N, Surges, R, and Duun-Henriksen, J. Expert perspective: who may benefit most from the new ultra long-term subcutaneous EEG monitoring? Front Neurol. (2021) 12:817733
78. Stirling, RE, Maturana, MI, Karoly, PJ, Nurse, ES, McCutcheon, K, Grayden, DB, et al. Seizure forecasting using a novel sub-scalp ultra-long term EEG monitoring system. Front Neurol. (2021) 12:713794. doi: 10.3389/fneur.2021.713794
79. Remvig, LS, Duun-Henriksen, J, Fürbass, F, Hartmann, M, Viana, PF, Kappel Overby, AM, et al. Detecting temporal lobe seizures in ultra long-term subcutaneous EEG using algorithm-based data reduction. Clin Neurophysiol. (2022) 142:86–93. doi: 10.1016/j.clinph.2022.07.504
80. Viana, PF, Remvig, LS, Duun-Henriksen, J, Glasstetter, M, Dümpelmann, M, Nurse, ES, et al. Signal quality and power spectrum analysis of remote ultra long-term subcutaneous EEG. Epilepsia. (2021) 62:1820–8. doi: 10.1111/epi.16969
81. Ulate-Campos, A, and Loddenkemper, T. Review on the current long-term, limited lead electroencephalograms. Epilepsy Behav. (2024) 150:109557. doi: 10.1016/j.yebeh.2023.109557
82. Lehnen, J, Venkatesh, P, Yao, Z, Aziz, A, Nguyen, PVP, Harvey, J, et al. Real-time seizure detection using behind-the-ear wearable system. J Clin Neurophysiol. (2024). doi: 10.1097/WNP.0000000000001076
83. Joyner, M, Hsu, SH, Martin, S, Dwyer, J, Chen, DF, Sameni, R, et al. Using a standalone ear-EEG device for focal-onset seizure detection. Bioelectron Med. (2024) 10:4. doi: 10.1186/s42234-023-00135-0
84. Munch Nielsen, J, Zibrandtsen, IC, Masulli, P, Lykke Sørensen, T, Andersen, TS, and Wesenberg Kjær, T. Towards a wearable multi-modal seizure detection system in epilepsy: a pilot study. Clin Neurophysiol. (2022) 136:40–8. doi: 10.1016/j.clinph.2022.01.005
85. Zeydabadinezhad, M, Jowers, J, Buhl, D, Cabaniss, B, and Mahmoudi, B. A personalized earbud for non-invasive long-term EEG monitoring. J Neural Eng. (2024) 21:026026. doi: 10.1088/1741-2552/ad33af
86. Bernini, A, Dan, J, and Ryvlin, P. Ambulatory seizure detection. Curr Opin Neurol. (2024) 37:99–104. doi: 10.1097/WCO.0000000000001248
87. Seth, EA, Watterson, J, Xie, J, Arulsamy, A, Md Yusof, HH, Ngadimon, IW, et al. Feasibility of cardiac-based seizure detection and prediction: a systematic review of non-invasive wearable sensor-based studies. Epilepsia Open. (2024) 9:41–59. doi: 10.1002/epi4.12854
88. Meritam Larsen, P, and Beniczky, S. Non-electroencephalogram-based seizure detection devices: state of the art and future perspectives. Epilepsy Behav. (2023) 148:109486. doi: 10.1016/j.yebeh.2023.109486
89. Mason, F, Scarabello, A, Taruffi, L, Pasini, E, Calandra-Buonaura, G, Vignatelli, L, et al. Heart rate variability as a tool for seizure prediction: a scoping review. J Clin Med. (2024) 13:30474. doi: 10.3390/jcm13030747
90. Komal, K, Cleary, F, Wells, JSG, and Bennett, L. A systematic review of the literature reporting on remote monitoring epileptic seizure detection devices. Epilepsy Res. (2024) 201:107334. doi: 10.1016/j.eplepsyres.2024.107334
91. Adams, T, Wagner, S, Baldinger, M, Zellhuber, I, Weber, M, Nass, D, et al. Accurate detection of heart rate using in-ear photoplethysmography in a clinical setting. Front Digit Health. (2022) 4:909519. doi: 10.3389/fdgth.2022.909519
92. Naganur, V, Sivathamboo, S, Chen, Z, Kusmakar, S, Antonic-Baker, A, O’Brien, TJ, et al. Automated seizure detection with noninvasive wearable devices: a systematic review and meta-analysis. Epilepsia. (2022) 63:1930–41. doi: 10.1111/epi.17297
93. Naganur, VD, Kusmakar, S, Chen, Z, Palaniswami, MS, Kwan, P, and O'Brien, TJ. The utility of an automated and ambulatory device for detecting and differentiating epileptic and psychogenic non-epileptic seizures. Epilepsia Open. (2019) 4:309–17. doi: 10.1002/epi4.12327
94. Kusmakar, S, Karmakar, CK, Yan, B, OrBrien, T, Palaniswami, M, and Muthuganapathy, R. Improved detection and classification of convulsive epileptic and psychogenic non-epileptic seizures using FLDA and Bayesian inference. Annu Int Conf IEEE Eng Med Biol Soc. (2018) 2018:3402–5. doi: 10.1109/EMBC.2018.8512981
95. Arends, J, Thijs, RD, Gutter, T, Ungureanu, C, Cluitmans, P, van Dijk, J, et al. Multimodal nocturnal seizure detection in a residential care setting: a long-term prospective trial. Neurology. (2018) 91:e2010–9. doi: 10.1212/WNL.0000000000006545
96. Velez, M, Fisher, RS, Bartlett, V, and le, S. Tracking generalized tonic-clonic seizures with a wrist accelerometer linked to an online database. Seizure. (2016) 39:13–8. doi: 10.1016/j.seizure.2016.04.009
97. Karoly, PJ, Stirling, RE, Freestone, DR, Nurse, ES, Maturana, MI, Halliday, AJ, et al. Multiday cycles of heart rate are associated with seizure likelihood: An observational cohort study. EBioMedicine. (2021) 72:103619. doi: 10.1016/j.ebiom.2021.103619
98. Goldenholz, DM, Goldenholz, SR, Romero, J, Moss, R, Sun, H, and Westover, B. Development and validation of forecasting next reported seizure using e-diaries. Ann Neurol. (2020) 88:588–95. doi: 10.1002/ana.25812
99. Fisher, RS, Boas, WE, Blume, W, Elger, C, Genton, P, Lee, P, et al. Epileptic seizures and epilepsy: definitions proposed by the international league against epilepsy (ILAE) and the International Bureau for Epilepsy (IBE). Epilepsia. (2005) 46:470–2. doi: 10.1111/j.0013-9580.2005.66104.x
100. Kanner, AM, Schachter, SC, Barry, JJ, Hesdorffer, DC, Mula, M, Trimble, M, et al. Depression and epilepsy: epidemiologic and neurobiologic perspectives that may explain their high comorbid occurrence. Epilepsy Behav. (2012) 24:156–68. doi: 10.1016/j.yebeh.2012.01.007
101. Gasparini, S, Ferlazzo, E, Leonardi, CG, Cianci, V, Mumoli, L, Sueri, C, et al. The natural history of epilepsy in 163 untreated patients: looking for “Oligoepilepsy”. PLoS One. (2016) 11:e0161722. doi: 10.1371/journal.pone.0161722
102. Goldenholz, DM, Eccleston, C, Moss, R, and Westover, MB. Prospective validation of a seizure diary forecasting falls short. Epilepsia. (2024). doi: 10.1111/epi.17984
103. Karoly, PJ, Goldenholz, DM, Freestone, DR, Moss, RE, Grayden, DB, Theodore, WH, et al. Circadian and circaseptan rhythms in human epilepsy: a retrospective cohort study. Lancet Neurol. (2018) 17:977–85. doi: 10.1016/S1474-4422(18)30274-6
104. Karoly, PJ, Freestone, DR, Boston, R, Grayden, DB, Himes, D, Leyde, K, et al. Interictal spikes and epileptic seizures: their relationship and underlying rhythmicity. Brain. (2016) 139:1066–78. doi: 10.1093/brain/aww019
105. Rosenow, F, and Luders, H. Presurgical evaluation of epilepsy. Brain. (2001) 124:1683–700. doi: 10.1093/brain/124.9.1683
106. Fotedar, N, Gajera, P, Pyatka, N, Nasralla, S, Kubota, T, Vaca, GFB, et al. A descriptive study of eye and head movements in versive seizures. Seizure. (2022) 98:44–50. doi: 10.1016/j.seizure.2022.04.003
107. Chiang, S, Khambhati, AN, Wang, ET, Vannucci, M, Chang, EF, and Rao, VR. Evidence of state-dependence in the effectiveness of responsive neurostimulation for seizure modulation. Brain Stimul. (2021) 14:366–75. doi: 10.1016/j.brs.2021.01.023
108. Karoly, PJ, Rao, VR, Gregg, NM, Worrell, GA, Bernard, C, Cook, MJ, et al. Cycles in epilepsy. Nat Rev Neurol. (2021) 17:267–84. doi: 10.1038/s41582-021-00464-1
109. Hofstra, WA, and de Weerd, AW. The circadian rhythm and its interaction with human epilepsy: a review of literature. Sleep Med Rev. (2009) 13:413–20. doi: 10.1016/j.smrv.2009.01.002
110. Schulze-Bonhage, A, Richardson, MP, Brandt, A, Zabler, N, Dümpelmann, M, and San Antonio-Arce, V. Cyclical underreporting of seizures in patient-based seizure documentation. Ann Clin Transl Neurol. (2023) 10:1863–72. doi: 10.1002/acn3.51880
111. Karoly, PJ, Freestone, DR, Eden, D, Stirling, RE, Li, L, Vianna, PF, et al. Epileptic seizure cycles: six common clinical misconceptions. Front Neurol. (2021) 12:720328. doi: 10.3389/fneur.2021.720328
112. Biagini, G, Panuccio, G, and Avoli, M. Neurosteroids and epilepsy. Curr Opin Neurol. (2010) 23:170–6. doi: 10.1097/WCO.0b013e32833735cf
113. Feely, M, Calvert, R, and Gibson, J. Clobazam in catamenial epilepsy. A model for evaluating anticonvulsants. Lancet. (1982) 320:71–3. doi: 10.1016/S0140-6736(82)91691-9
114. Saboo, KV, Cao, Y, Kremen, V, Sladky, V, Gregg, NM, Arnold, PM, et al. Individualized seizure cluster prediction using machine learning and chronic ambulatory intracranial EEG. IEEE Trans Nanobioscience. (2023) 22:818–27. doi: 10.1109/TNB.2023.3275037
115. Stribling, D, Xia, Y, Amer, MK, Graim, KS, Mulligan, CJ, and Renne, R. The model student: GPT-4 performance on graduate biomedical science exams. Sci Rep. (2024) 14:5670. doi: 10.1038/s41598-024-55568-7
116. Luke, W, Seow Chong, L, Ban, KH, Wong, AH, Zhi Xiong, C, Shuh Shing, L, et al. Is ChatGPT ready to be a learning tool for medical undergraduates and will it perform equally in different subjects? Comparative study of ChatGPT performance in tutorial and case-based learning questions in physiology and biochemistry. Med Teach. (2024):1–7. doi: 10.1080/0142159X.2024.2308779
117. Kerr, WT, Chen, H, Figuera Losada, M, Cheng, C, Liu, T, and French, J. Reasons for ineligibility for clinical trials of patients with medication resistant epilepsy. Epilepsia. (2023) 64:e56–60. doi: 10.1111/epi.17568
118. Ryvlin, P, Cucherat, M, and Rheims, S. Risk of sudden unexpected death in epilepsy in patients given adjunctive antiepileptic treatment for refractory seizures: a meta-analysis of placebo-controlled randomised trials. Lancet Neurol. (2011) 10:961–8. doi: 10.1016/S1474-4422(11)70193-4
119. Goldenholz, DM, Goldenholz, EB, and Kaptchuk, TJ. Quantifying and controlling the impact of regression to the mean on randomized controlled trials in epilepsy. Epilepsia. (2023) 64:2635–43. doi: 10.1111/epi.17730
120. French, JA, Gil-Nagel, A, Malerba, S, Kramer, L, Kumar, D, and Bagiella, E. Time to prerandomization monthly seizure count in perampanel trials: a novel epilepsy endpoint. Neurology. (2015) 84:2014–20. doi: 10.1212/WNL.0000000000001585
121. Hannon, T, Fernandes, KM, Wong, V, Nurse, ES, and Cook, MJ. Over-and underreporting of seizures: how big is the problem? Epilepsia. (2024) 65:1406–14. doi: 10.1111/epi.17930
Keywords: epilepsy, wearables, internet of things, human-in-the loop, deficiency time
Citation: Kerr WT, McFarlane KN and Figueiredo Pucci G (2024) The present and future of seizure detection, prediction, and forecasting with machine learning, including the future impact on clinical trials. Front. Neurol. 15:1425490. doi: 10.3389/fneur.2024.1425490
Edited by:
James Hillis, Harvard Medical School, United StatesReviewed by:
Attila Racz, University Hospital Bonn, GermanyKapil Gururangan, University of California, Los Angeles, United States
Copyright © 2024 Kerr, McFarlane and Figueiredo Pucci. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Wesley T. Kerr, S2VycldAUGl0dC5lZHU=