- 1Institute of Psychology, University of the Bundeswehr Munich, Neubiberg, Germany
- 2Department of Psychiatry and Psychotherapy, University of Regensburg, Regensburg, Germany
- 3Department of Electrical Engineering, University of the Bundeswehr Munich, Neubiberg, Germany
- 4Institute of Sport Science, University of the Bundeswehr Munich, Neubiberg, Germany
Introduction: Repetitive transcranial magnetic stimulation (rTMS) is used to induce long-lasting changes (aftereffects) in cortical excitability, which are often measured via single-pulse TMS (spTMS) over the motor cortex eliciting motor-evoked potentials (MEPs). rTMS includes various protocols, such as theta-burst stimulation (TBS), paired associative stimulation (PAS), and continuous rTMS with a fixed frequency. Nevertheless, subsequent aftereffects of rTMS are variable and seem to fail repeatability. We aimed to summarize standard rTMS procedures regarding their test–retest reliability. Hereby, we considered influencing factors such as the methodological quality of experiments and publication bias.
Methods: We conducted a literature search via PubMed in March 2023. The inclusion criteria were the application of rTMS, TBS, or PAS at least twice over the motor cortex of healthy subjects with measurements of MEPs via spTMS as a dependent variable. The exclusion criteria were measurements derived from the non-stimulated hemisphere, of non-hand muscles, and by electroencephalography only. We extracted test–retest reliability measures and aftereffects from the eligible studies. With the Rosenthal fail-safe N, funnel plot, and asymmetry test, we examined the publication bias and accounted for influential factors such as the methodological quality of experiments measured with a standardized checklist.
Results: A total of 15 studies that investigated test–retest reliability of rTMS protocols in a total of 291 subjects were identified. Reliability measures, i.e., Pearson's r and intraclass correlation coefficient (ICC) applicable from nine studies, were mainly in the small to moderate range with two experiments indicating good reliability of 20 Hz rTMS (r = 0.543) and iTBS (r = 0.55). The aftereffects of rTMS procedures seem to follow the heuristics of respective inhibition or facilitation, depending on the protocols' frequency, and application pattern. There was no indication of publication bias and the influence of methodological quality or other factors on the reliability of rTMS.
Conclusion: The reliability of rTMS appears to be in the small to moderate range overall. Due to a limited number of studies reporting test–retest reliability values and heterogeneity of dependent measures, we could not provide generalizable results. We could not identify any protocol as superior to the others.
1. Introduction
Transcranial magnetic stimulation (TMS) is a non-invasive brain stimulation (NIBS) technique with which brain activity can be induced and modulated. By applying TMS via single pulses (spTMS), momentary states of cortical excitability can be assessed; for example, stimulations over the motor cortex can induce a motor-evoked potential (MEP) in the respective contralateral hand muscle (Rossini and Rossi, 1998). TMS applied in a repetitive manner (rTMS) is deployed with either a fixed frequency with common frequencies of 1, 10, or 20 Hz or with complex patterns to induce longer-lasting neuroplastic changes (aftereffects) in the brain (Hallett, 2000; Siebner and Rothwell, 2003). MEPs are also used to capture these aftereffects qualitatively. Literature suggests that changes in cortical excitability after rTMS are depending on stimulation frequency (Fitzgerald et al., 2006). Hereby, frequencies at approximately 1 Hz lead to the inhibition of neuronal activity, i.e., lower MEPs after rTMS than before, and stimulation with frequencies over 5 Hz evoke facilitatory aftereffects, i.e., higher MEPs after rTMS than at the baseline. This assumption is referred to as the low-frequency inhibitory–high-frequency excitatory (lofi-hife) heuristic (Prei et al., 2023). Common patterned rTMS procedures are paired associative stimulation (PAS), whereby electrical stimulation of the respective peripheral muscle (conditioning stimulation; CS) is applied in close relation to the contralateral TMS pulse, and theta-burst stimulation (TBS), whereby triplets at 50 Hz are repeatedly delivered with a 5-Hz inter-burst-interval. The latter can be administered as intermittent TBS (iTBS) with a pause of 8 s between 2 s of stimulation or as continuous TBS (cTBS) without breaks. Regarding aftereffects, cTBS is considered to elicit inhibitory effects as well as PAS with an inter-stimulus interval (ISI) of 10 ms between CS and TMS pulse (PAS10), whereas iTBS and PAS with an ISI of 25 ms (PAS25) tend to be excitatory (Huang et al., 2005; Wischnewski and Schutter, 2016). All of these procedures have been frequently used in both basic and clinical research settings as well as in the treatment of various neurological and psychiatric disorders (Berlim et al., 2013, 2017; Patel et al., 2020; Shulga et al., 2021). However, recent literature shows high inter- and intra-subject variability in rTMS aftereffects, questioning the heuristics of a clear association of inhibition or facilitation with a specific protocol (Fitzgerald et al., 2006).
The variability of TMS and rTMS outcome parameters has been a major topic in the NIBS community (Guerra et al., 2020; Goldsworthy et al., 2021). Furthermore, with high variability of rTMS aftereffects, their reliability can decrease. As one of the three main quality criteria of scientific experiments, test–retest reliability is important because it indicates whether a measurement or intervention is precise and can be repeated over time while generating the same output. Reliability is most commonly measured with Pearson's r (r) and intraclass correlation coefficient (ICC), but both measures are not always reported. Moreover, there is no consensus on whether rTMS reliability is assessed for the measurements of MEPs after rTMS only (post) or rTMS aftereffects (MEPs from before rTMS application subtracted from MEPs after rTMS). To date, the focus of rTMS research has been primarily on the identification and enhancement of aftereffects, but whether these effects are reproducible over time has been rather neglected. With this review, we aimed to give an overview and classification of test–retest reliability of rTMS procedures. We investigated whether there is a most effective protocol to reliably induce neuroplastic changes in the brain. We conducted a meta-analysis with moderator variables to exclude that study-inherent parameters influence the reliability outcome. Hereby, we refer to parameters that can influence aftereffects of NIBS and their variability and depend on the equipment and schedule of laboratories and that cannot be adjusted throughout the experiment. These parameters we wanted to account for are the use of neuronavigation (Herwig et al., 2001; Julkunen et al., 2009), sex of participants (Pitcher et al., 2003), repetition interval (Hermsen et al., 2016), year of publication, excitability of protocol, and methodological quality of the study. To assess methodological quality, we chose the checklist by Chipchase et al. (2012). The authors created a checklist via the Delphi procedure that assesses whether publications descriptively “report” and experimentally “control” for participant factors, methodological factors, and analytical factors, which are observed as likely to influence MEP responses elicited by TMS. With the percentage of applicable items from the checklist, one has an approximate measure of the overall methodological quality of a study ranging from 0 to 100%. To validate our assessment, we checked for interrater agreement. Moreover, we have provided an overview of rTMS aftereffects for each of the protocols within the studies assessing reliability. In order to identify the most effective protocol to induce neuroplastic changes, we compared these protocol-specific aftereffects. Publication bias was examined to assess whether the reliability values of current studies are representative.
2. Methods
2.1. Inclusion criteria
Author CK performed a literature search in PubMed (latest in March 2023) using the keywords (“rTMS” OR “TBS” OR “theta burst” OR “PAS” OR paired asso* stim* OR repet* transc* magn* stim*) AND (MEP OR motor evoked potentials OR cort* exci* OR plast*) AND (reli* OR reproduc* OR repeat*). No filters or automation tools were applied. Based on the PRISMA flow diagram schema (Page et al., 2021), we extracted articles with experiments that applied (1) any kind of rTMS, TBS, or PAS (2) the technique at least twice (3) stimulation area over the motor cortex (4) investigating healthy subjects (5) with measurements of MEPs via spTMS as a dependent variable. As target muscles, we focused on hand muscles to ensure comparability. In the first step, author CK screened records in the PubMed database by title and abstract. Second, authors CK and MO inspected independently the full text for the abstracts screened to confirm eligibility and inclusion in the review. The exclusion criteria were (1) measures derived from the non-stimulated hemisphere, i.e., different stimulation locations of rTMS and spTMS, (2) examination of non-hand muscles only, and (3) reports of only electroencephalography measurements. Other search strings did not manage to find all the articles we had included.
2.2. Test–retest reliability
For interpreting reliability and conducting the meta-analysis, we only used the studies in which a reliability measure such as ICC or r was conducted. Moreover, whenever two dependent measures of the identical session were made or both ICC and r values were calculated, we only used one value each for depiction and within the meta-analysis in order to prevent overrepresentation. Preferably, we extracted the r value over the whole measurement because it has a fixed range of values, whereas ICC calculations need to be correctly selected and reported to exclude biases. Nevertheless, in the summarizing Table 1, all reliability values are listed.
We compiled an overview of reliability values of the different rTMS protocols and interpreted r after Cohen (1988), with r < 0.1 representing the very small range, 0.1 ≤ r < 0.3 small, 0.3 ≤ r < 0.5 medium, and 0.5 ≤ r ≤ 1 the large range. ICCs were interpreted after Koo and Li (2016), with ICC of < 0.5 being poor, 0.5 ≤ ICC < 0.75 being moderate, 0.75 ≤ ICC < 0.9 being good, and ICC ≥ 0.9 being excellent. To assess whether one rTMS protocol might be superior in reliability to others, we conducted the Kruskal–Wallis test.
2.3. Influences of rTMS reliability
In a chi-squared test by Hunter and Schmidt (2000), we assessed the homogeneity of reliability values. To identify whether study-inherent parameters influence rTMS reliability measures, we conducted a random effects regression analysis with Fisher's z-transformed reliability values using the following continuous predictors: “methodological quality”, “year of publication”, “repetition interval of rTMS”, and “sex ratio” as well as categorical predictors “neuronavigation” and “excitability of the protocol”. Hereby, two authors (CK and MO) assessed the methodological quality via the checklist from Chipchase et al. (2012). To ensure the objectivity of the procedure, we calculated Cohen's kappa (κ) for interrater agreement per study with confidence intervals (CI) (Cohen, 1960). A detailed description of how we applied the checklist can be found in the Supplementary material. To assess the influence of the participants' sex, for each study, the sex ratio of the sample was the ratio of the number of female participants per male participants.
To investigate publication bias, we calculated the Rosenthal fail-safe N (Rosenthal, 1979) and conducted a funnel plot for Fisher's z-transformed reliability values and the respective standard errors derived from the sample sizes. By testing the funnel plot for asymmetry (after Egger et al., 1997), we could assess the influence of publication bias on rTMS reliability. We correlated reliability values with the respective year of publication to assess whether more recent research could generate higher reliability of rTMS.
Analyses were run in SPSS (IBM Corp., Version 29) and R (R Core Team, Austria, Version 4.0.5) with the meta package (version 6.2-1) (Schwarzer, 2007). Corresponding syntaxes for the meta-analysis could be found by Field and Gillett (2010).
2.4. Aftereffects of rTMS protocols of included studies
We summarized the aftereffects of rTMS, namely, the percent change from baseline MEP to MEP after rTMS stimulation. With the Kruskal–Wallis test, we tested if rTMS protocols had comparable aftereffects. For those studies that did not provide a mean percent change measure, we recalculated mean changes with the provided descriptive data in the manuscripts.
3. Results
3.1. Study selection and characteristics
A total of 819 articles resulted from the search in PubMed (latest in March 2023). One preprint was co-authored by CK and included into consideration. In total, 15 articles met the inclusion criteria and were identified as eligible for this review. The detailed procedure can be retraced via the PRISMA flow chart depicted in Figure 1.
Figure 1. PRISMA workflow of how the articles were identified in PubMed via search string and the procedure of identifying the included studies in this review (Page et al., 2021).
The included studies applied continuous rTMS protocols of 1 Hz rTMS (Maeda et al., 2000; Bäumer et al., 2003; Modugno et al., 2003; Prei et al., 2023), 2 Hz rTMS (Sommer et al., 2002), 10 Hz rTMS (Maeda et al., 2000), and 20 Hz rTMS (Maeda et al., 2000; Cohen et al., 2010; Prei et al., 2023). Moreover, test–retest reliability assessments of patterned rTMS protocols were found: iTBS (Hinder et al., 2014; Schilberg et al., 2017; Perellón-Alfonso et al., 2018; Boucher et al., 2021), cTBS (Vernet et al., 2014; Vallence et al., 2015; Jannati et al., 2019; Boucher et al., 2021), and PAS25 (Fratello et al., 2006; Sale et al., 2007). Two of these studies included sham stimulation (Perellón-Alfonso et al., 2018; Boucher et al., 2021).
The mean sample size of studies was n = 17 (range 4–30, total: 291), with overall 138 women, 143 men, and 10 not applicable. Participants had a mean age of 27 years (range: 18–65 years). The mean test–retest intervals were 13.41 days (range: 6 h to 107 days).
Fratello et al. (2006) reported ICCs for the muscle at which the representative cortical spot was stimulated, i.e., abductor pollicis brevis (APB), as well as one other muscle, the abductor digiti minimi (ADM). For both muscles, the authors computed ICCs with the respective post-rTMS measure as well as with the rTMS aftereffects (post-pre). Sale et al. (2007) computed ICCs for a long and a short application of PAS25 as well as for the groups that attended sessions in the morning and afternoon, respectively. Boucher et al. (2021) reported r values for the overall measurement of iTBS, cTBS, sham, and ICCs for 5, 10, 20, 30, 50, and 60 min after rTMS application. Hereby, we summarized the ICCs by reporting the minimal and maximal ICC of the respective procedures. Jannati et al. (2019) assessed the reliability values from the area under the curve (AUC) of elicited MEPs as well as for the area under the curve of rectified MEPs (|AUC|). Vallence et al. (2015) reported r values for two sessions each, i.e., sessions 1 and 2 as well as sessions 1 and 3. The authors also conducted ICC for rTMS aftereffect assessment at stimulus intensities that elicited peak-to-peak MEP amplitudes of 1 mV ± 0.15 mV (SI1 mV). Additionally, ICC values for stimulus intensities measured via the input–output curve (IO curve) that evoked MEP sizes halfway between minimal and maximal cortical excitability (SI50), supra-threshold stimulus intensities (150% RMT), and stimulus intensities evoking maximal MEPs (180% RMT) were conducted. Prei et al. (2023) reported r as well as ICCs of mean and median MEPs over the whole measurement and the respective quarters, whereby we report the overall measures. Table 1 gives an overview of the methodological and result parameters of interest.
3.2. Test–retest reliability
Five of the 15 studies did not report any reliability value but only measures of variance (Sommer et al., 2002; Bäumer et al., 2003; Modugno et al., 2003; Cohen et al., 2010; Vernet et al., 2014). One study only conducted ICCs for baseline MEPs but not for MEPs after the rTMS procedure (Perellón-Alfonso et al., 2018). Therefore, these six studies were excluded from the subsequent analyses of reliability, resulting in nine evaluable studies. Both ICC and r values were reported in five studies; one publication conducted only r and three only ICCs. We preferably extracted r values instead of ICC as well as the overall reliability for a respective protocol within the studies. Whenever reliability values from the same sessions were calculated, we chose mean MEP amplitudes instead of median (Prei et al., 2023) and values derived from the muscle whose cortical representation was stimulated (Fratello et al., 2006) as well as the non-rectified derived parameter (Jannati et al., 2019). Five out of nine studies reported the reliability of aftereffect measures, one reported the reliability of post-rTMS measures (Hinder et al., 2014), two did not state which measure they chose (Sale et al., 2008; Boucher et al., 2021), and one reported both the post and the post-pre measures (Fratello et al., 2006). The highest value was used (post-measure) obtained from the study of Fratello et al. (2006) for depiction and meta-analysis in order to gain the most information out of the publication bias investigation.
Figure 2 gives an overview of the included comparable values from evaluable studies combined with a representation of the number of participants (sample size) per experiment and the time interval of rTMS repetition. ICCs in this overview of reliability measures ranged from 0.29 to 0.7 and r from 0.097 to 0.55 with one iTBS reliability value being in the negative range (−0.284). Reliabilities did not differ between rTMS protocols in analysis via the Kruskal–Wallis test ( = 5.292, p = 0.381). The exact values can be found in Table 1. According to the classification after Cohen (1988), rTMS reliability yields small to medium effect sizes. The reliabilities of the 20 Hz rTMS protocol (Maeda et al., 2000) and one iTBS protocol (Hinder et al., 2014) show large effect sizes (r = 0.543 and r = 0.55, respectively). ICCs also range from poor to moderate reliability values (Koo and Li, 2016). 9 out of 15 rTMS reliability values were interpreted as small or poor, 4 out of 15 as medium or moderate and 2 out of 15 as large effects sizes.
Figure 2. Bar graph of Pearson's r (upper) and ICC values (lower) of the rTMS protocols from the respective articles identified as eligible and comparable in this review. Reliability values are sorted by protocol and the respective excitability. Patterns of the bars represent in which article the respective values were published. The width of the bars represents the sample size within the respective studies. On the right side of the bars, the (mean or minimum of) number of days between repetition of the rTMS protocol (repetition interval; ri) are depicted. Colors represent the interpretation of r (Cohen, 1988) and ICC (Koo and Li, 2016) values as negligible (darker red), poor/low (lighter red), medium/moderate (yellow), good/large (lighter green), or excellent (darker green).
3.3. Influences of rTMS reliability
All analyses include both r and ICC values because otherwise not all rTMS protocols would be covered. No significant heterogeneity between study reliabilities was found via the chi-squared test ( = 20.945, p = 0.103). The random effects regression analysis revealed an overall mean Fisher's z-transformed reliability value of 0.315 (95% CI range: 0.168 to 0.449) that is significantly different from zero (t(15) = 4.45, p = 0.001) in a model without any predictors. By including the continuous and categorical predictors, none of them showed a significant influence (continuous: t(7) < 1, p > 0.442 and categorical: < 1, p > 0.714). Methodological quality assessed with the checklist by Chipchase et al. (2012) ranged from 30.8 to 55.8% and interrater agreement from 0.65 (CI range: 0.5 to 0.82) to 1 (CI range: 1 to 1). Detailed ratings can be found in Supplementary Tables S1, S2. Additionally, the goodness of fit of the random effects regression analysis was not significant ( = 7.935, p = 0.44). The Rosenthal fail-safe N was 164. The funnel plot, representing Fisher's z-transformed reliability values on the x-axis and the standard errors from the random effects model on the y-axis, is depicted in Figure 3. With a test for funnel plot asymmetry (after Egger et al., 1997), no significant skewedness was identified (t(13) = 1.42, p = 0.179). Correlating reliability values with their respective year of publication revealed a negative but non-significant result (r = −0.475, p = 0.073), which is shown in Figure 4.
Figure 3. Funnel plot of the standardized effect sizes (Fisher z-transformed values of r and ICC) of the eligible studies plotted against their standard error from the random effects model (circles). The diagonal lines (gray background) depict the 95% confidence interval.
Figure 4. Reliability values as black dots included in the meta-analysis with the year of publication of the respective studies in a scatter plot. The negative trend from the correlation analysis is depicted as the black line.
3.4. Aftereffects of rTMS procedures of included studies
From four experiments (Sommer et al., 2002; Bäumer et al., 2003; Vallence et al., 2015; Prei et al., 2023), we were not able to retrieve information on aftereffect sizes, resulting in 11 studies whose data can be included in a summary of aftereffects (for values see Table 1). The size and direction of aftereffects for each rTMS protocol are depicted in Figure 5. The protocols known to primarily evoke cortical inhibition (cTBS and 1 Hz) mainly show the inhibitory effects of MEP amplitudes after rTMS application. Sham stimulation has both aftereffect directions. Excitatory protocols (20 Hz, iTBS, PAS25) mostly evoke cortical excitation, except for the 10 Hz protocol that has inhibitory tendencies. Analysis via the Kruskal–Wallis test revealed differences in aftereffects between the rTMS protocols ( = 20.762, p = 0.002), whereby cTBS had significantly lower effects than 20 Hz rTMS (p = 0.005), iTBS (p < 0.001), and PAS25 (p = 0.001), and 1 Hz rTMS had significantly lower aftereffects than 20 Hz rTMS (p = 0.033), iTBS (p = 0.008), and PAS25 (p = 0.008). Nevertheless, within excitatory or inhibitory protocols, there were no differences in aftereffects (p > 0.054).
Figure 5. Bar graph of aftereffect sizes, i.e., the percent change of MEPs after rTMS relative to pre-rTMS (baseline) MEPs of the data collected from the eligible studies and sorted for each rTMS protocol. Protocols are grouped for inhibitory and excitatory effect expectations according to the TMS-protocol-dependent inhibition-excitation heuristic. Positive aftereffect sizes indicate cortical excitation after rTMS stimulation compared to baseline, whereas negative aftereffect sizes refer to cortical inhibition.
4. Discussion
This study provides an overview of test–retest reliability of rTMS procedures. We assessed whether we could identify a procedure to be most effective and reliably induce neuroplastic changes in the brain. We identified whether parameters inherent to the respective publication influenced reliability values. To estimate the representativity of reliability values, we checked for publication bias.
A total of 15 studies were found that assessed whether rTMS protocols evoke repeatable cortical reactions over time, i.e., their test–retest reliability. Pearson's r and ICC values were interpreted as small in nine and medium in four studies. Only the reliabilities of the 20 Hz rTMS protocol (Maeda et al., 2000) and of one iTBS protocol (Hinder et al., 2014) are to be interpreted as large according to Cohen (1988). Nevertheless, there is no evidence to favor one particular rTMS protocol based on higher reliability values since no value showed a significant difference from the mean reliability in the Kruskal–Wallis test. Moreover, the chi-squared test did not confirm a difference between reliabilities, which additionally approves the joint analysis of Pearson's r and ICC values. One negative reliability value (r = −0.284) for iTBS was identified, meaning that the participants' first application of iTBS resulted in a facilitatory effect, whereas the second application had an inhibitory effect and vice versa (Boucher et al., 2021). The two other experiments assessing iTBS reliability did not show such a relation.
Parameters that depend on the publication, i.e., methodological quality of the study and year of publication, and on experimental design, i.e., excitability of protocol, use of neuronavigation, sex ratio of the sample, and rTMS repetition interval, had no influence on the reliability of rTMS. Publication bias did not affect the present reliability values indicated by the funnel plot and asymmetry test, which thus strengthens the credibility of these reliability estimates. Nevertheless, the funnel plot depicts that all reliability values are associated with relatively high standard errors, which are derived from the sample sizes of the studies. The Rosenthal fail-safe N of 164 indicates that, if 164 unpublished studies had a non-significant reliability parameter, the estimated overall reliability would turn significantly different from zero to non-significant. By correlating reliability values with the respective year of publication, we identified a negative trend, indicating that, in the current research study, reliability values decrease. This could be either due to higher publication bias in past publications or due to increasing variability perhaps induced by the plurality of setup equipment and higher accuracy of measures, e.g., TMS stimulators, coils, and electrodes. Moreover, higher objectivity of assessment might lead to reduced reliability values. Thus, further studies are needed with higher sample sizes and a systematic investigation of reliability to strengthen the assumption that no publication bias is present in rTMS reliability studies.
rTMS protocols of our included studies followed the lofi-hife heuristic and association with respective inhibition or facilitation effects. Most of the protocols mainly evoking inhibitory neuronal effects resulted in reduced MEPs after stimulation, and the protocols primarily having excitatory effects produced higher MEPs after stimulation. Only the 10 Hz rTMS protocol (Maeda et al., 2000) shows contrary results. The Kruskal–Wallis test confirmed that inhibitory rTMS protocols had significantly lower aftereffects than excitatory protocols, whereby 10 Hz rTMS did not differ significantly from both inhibitory and excitatory protocols.
Identification of the rTMS protocol with the most reliable and effective outcome cannot be provided currently. Since reliability values did not differ between protocols and also within excitatory and inhibitory protocols, aftereffects were comparable, no superiority of certain protocols can be proven. Descriptively, cTBS seems to have better inhibitory aftereffects and reliability than 1 Hz rTMS. For excitatory protocols, iTBS can induce descriptively higher aftereffects than PAS25, 20 Hz and 10 Hz rTMS, but also varies more in reliability.
In contrast to baseline spTMS test–retest reliability values reaching ICCs of 0.86 (Pellegrini et al., 2018), rTMS reliability values tend to be smaller. It is important to note that, to assess rTMS reliability, both rTMS and spTMS need to be applied. Thus, variability in both measures adds up to the resulting rTMS aftereffect reliability value. To assess cortical excitability, IO curves cover the whole spectrum best. Nonetheless, to gain clear insights into brain functions and effective treatment of disorders by rTMS, reliable measurements are necessary at best with reliability values in the large range. On the one hand, this is achieved by identifying and eliminating or controlling parameters that influence the variability of spTMS and rTMS. On the other hand, personalization of applications can be an effective method (Schoisswohl et al., 2021).
Many parameters that influence the variability of spTMS and rTMS are already identified, e.g., stimulation intensity and number of applied pulses (Pascual-Leone et al., 1994; Fitzgerald et al., 2002; Peinemann et al., 2004; Lang et al., 2006), pulse form (Arai et al., 2005), time of day (Sale et al., 2008), subject-related factors such as age (Rossini et al., 1992; Pitcher et al., 2003; Todd et al., 2010; Cueva et al., 2016), genetic factors (Cheeran et al., 2008; Di Lazzaro et al., 2015), and changes in motor activation state (Huang et al., 2008; Iezzi et al., 2008; Goldsworthy et al., 2014). The identified studies assessing rTMS reliability show that, with higher stimulus intensities, cortical inhibition increases during cTBS and perceived stress correlated with larger aftereffects (Vallence et al., 2015). A 20 Hz rTMS application with a night in between resulted in higher aftereffects than stimulation overday (Cohen et al., 2010). Influences on the reliability of rTMS procedures were assessed by Jannati et al. (2019), who showed in an exploratory analysis that age and genotype had an influence on the reliability of cTBS aftereffects. Sale et al. (2007) revealed that PAS25 assessment in the afternoon is more reliable than in the morning. To systematically investigate which parameters affect the reliability of rTMS, further studies are needed with more power and randomized and controlled experimental design. A meta-analysis is hereby not sufficient to extract dependable information. To generate comparable data on rTMS reliability, future studies should report both Pearson's r value and ICC with corresponding confidence intervals as well as the model that the ICC calculation was based on. It should also be established to compute and report both reliability of post-rTMS measures and rTMS aftereffects. Further research on the variability and test–retest reliability of rTMS procedures is needed to identify factors that improve rTMS reliability and estimate the maximal reliability values achievable.
Although the induction of expected inhibitory or facilitatory aftereffects by rTMS protocols seems to succeed, high inter- and intra-individual variabilities dominate the results of rTMS experiments (Schilberg et al., 2017), even when controlling for most influencing parameters. In an experiment by Hamada et al. (2013), 50% of the aftereffect variation after TBS was predicted by a marker for late I-wave recruitment, which is discussed to be a mechanism of neuromodulation (Di Lazzaro et al., 2004). Thus, the other 50% of the variation is not yet explainable, also being not related to age, gender, time of day, and baseline MEP sizes (Hamada et al., 2013). It raises the question whether rTMS in general leads to the induction of variability in the neuronal responses and thus does not achieve exclusively LTP- or LTD-like plasticity effects. Thus, investigations of variability, e.g., the coefficient of variation in addition to the mean evoked responses could provide explanations. Additionally, all-encompassing sham conditions could reveal unbiased aftereffects.
Another approach to identifying reasons for the variability of rTMS is to investigate patients. For example, patients with Alzheimer's disease characterized by neuronal degeneration and rigidity show higher reliability of rTMS compared to healthy controls (Fried et al., 2017). One explanation might be that, in Alzheimer's patients, impairments of cortical plasticity can lead to omitted rTMS aftereffects (Di Lorenzo et al., 2020). Nevertheless, there is also evidence that patients with more severe Alzheimer's disease markers show higher inhibitory aftereffects after 1 Hz rTMS (Koch et al., 2011), which might indicate that neuronal rigidity can be altered by the induction of variability with rTMS. Thus, interpretive approaches should be taken with caution.
The present systematic review and meta-analysis address the test–retest reliability of rTMS on healthy individuals, and derived findings cannot be transferred to other populations, such as patient groups, or other applications, e.g., stimulation over other cortices than the motor cortex. Participants from the included studies were often right-handed and accordingly did not show a representative sample. Because only a few studies contributed to the analysis of reliability values, the results cannot be generalized and need to be interpreted with caution. Additionally, although eliciting MEPs is a common procedure to investigate cortical excitability, it still represents an indirect measure. There are other markers that hold the potential to estimate reactions of the brain to rTMS in a more direct way, yet, to date, they are studied less frequently and prone to artifacts and noise.
Test–retest reliability of rTMS in the identified studies is mainly small to moderate, with overall scarce experimental assessment. Aftereffects of rTMS protocols mainly followed the respective inhibition or excitation expectation. No protocol is to be favored based on our findings of reliability values and aftereffect sizes. However, the generalizability remains questionable because of limited comparable data. By reporting ICC as well as Pearson's r values of both post-rTMS and aftereffect measures, studies examining test–retest reliability can contribute to comparability. Additionally, the application of spTMS should be equal, e.g., by assessing the IO curves of MEPs. In general, the variability of NIBS outcomes is mirrored in its reliability. Influential factors of both spTMS and rTMS need to be systematically investigated to achieve high and reliable rTMS aftereffects. To establish rTMS procedures in the clinical everyday use of disorder treatment, higher reliability is necessary. With this overview, scientists and clinicians can estimate and compare the size and reliability of the aftereffects of rTMS based on current data.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
CK, SS, and MS contributed to the review concept and design. CK performed the literature search and conducted data analysis and visualized the study and drafted the manuscript. CK and MO reviewed included studies and examined the studies' methodological quality. CK, MO, FS, KL, WS, WM, SS, and MS revised the manuscript. All authors have read and approved the final version. All authors contributed to the article and approved the submitted version.
Funding
This research is funded by dtec.bw—Digitalization and Technology Research Center of the Bundeswehr (project MEXT). Dtec.bw is funded by the European Union—NextGenerationEU.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2023.1237713/full#supplementary-material
Abbreviations
APB, abductor pollicis brevis; AUC, area under the curve; |AUC|, area under the curve of rectified MEPs; CI, confidence interval; cTBS, continuous theta-burst stimulation; EMG, electromyography; FDI, first dorsal interosseous muscle; ICC, intraclass correlation coefficient; IO curve, input–output curve; ISI, inter-stimulus interval; iTBS, intermittent theta-burst stimulation; κ, Cohen's kappa; lofi-hife heuristic, a heuristic which refers to the assumption that continuous low-frequency repetitive transcranial magnetic stimulation leads to cortical inhibition, whereas continuous high-frequency repetitive transcranial magnetic stimulation leads to cortical facilitation; MEP, motor-evoked potential; NIBS, non-invasive brain stimulation; PAS, paired associative stimulation; PAS10, paired associative stimulation with 10 ms between TMS pulse and conditioning pulse; PAS25, paired associative stimulation with 25 ms between TMS pulse and conditioning pulse; r, Pearson's correlation coefficient; RMT, resting motor threshold; rTMS, repetitive transcranial magnetic stimulation; SI1mV, stimulus intensities that elicited peak-to-peak MEP amplitudes of 1 mV ± 0.15 mV; SI50, stimulus intensities that evoke MEP sizes halfway between minimal and maximal cortical excitability; spTMS, transcranial magnetic stimulation applied with single pulses; TBS, theta-burst stimulation; TMS, transcranial magnetic stimulation.
References
Arai, N., Okabe, S., Furubayashi, T., Terao, Y., Yuasa, K., and Ugawa, Y. (2005). Comparison between short train, monophasic and biphasic repetitive transcranial magnetic stimulation (rTMS) of the human motor cortex. Clin. Neurophysiol. 116, 605–613. doi: 10.1016/j.clinph.2004.09.020
Bäumer, T., Lange, R., Liepert, J., Weiller, C., Siebner, H. R., Rothwell, J. C., et al. (2003). Repeated premotor rTMS leads to cumulative plastic changes of motor cortex excitability in humans. Neuroimage 20, 550–560. doi: 10.1016/S1053-8119(03)00310-0
Berlim, M. T., McGirr, A., Rodrigues Dos Santos, N., Tremblay, S., and Martins, R. (2017). Efficacy of theta burst stimulation (TBS) for major depression: An exploratory meta-analysis of randomized and sham-controlled trials. J. Psychiatr. Res. 90, 102–109. doi: 10.1016/j.jpsychires.2017.02.015
Berlim, M. T., van den Eynde, F., and Daskalakis, Z. J. (2013). A systematic review and meta-analysis on the efficacy and acceptability of bilateral repetitive transcranial magnetic stimulation (rTMS) for treating major depression. Psychol. Med. 43, 2245–2254. doi: 10.1017/S0033291712002802
Boucher, P. O., Ozdemir, R. A., Momi, D., Burke, M. J., Jannati, A., Fried, P. J., et al. (2021). Sham-derived effects and the minimal reliability of theta burst stimulation. Sci. Rep. 11, 21170. doi: 10.1038/s41598-021-98751-w
Cheeran, B., Talelli, P., Mori, F., Koch, G., Suppa, A., Edwards, M., et al. (2008). A common polymorphism in the brain-derived neurotrophic factor gene (BDNF) modulates human cortical plasticity and the response to rTMS. J. Physiol. 586, 5717–5725. doi: 10.1113/jphysiol.2008.159905
Chipchase, L., Schabrun, S., Cohen, L., Hodges, P., Ridding, M., Rothwell, J., et al. (2012). A checklist for assessing the methodological quality of studies using transcranial magnetic stimulation to study the motor system: an international consensus study. Clin. Neurophysiol. 123, 1698–1704. doi: 10.1016/j.clinph.2012.05.003
Cohen, D. A., Freitas, C., Tormos, J. M., Oberman, L., Eldaief, M., and Pascual-Leone, A. (2010). Enhancing plasticity through repeated rTMS sessions: the benefits of a night of sleep. Clin. Neurophysiol. 121, 2159–2164. doi: 10.1016/j.clinph.2010.05.019
Cohen, J. (1960). A coefficient of agreement for nominal scales. Educ. Psychol. Measur. 20, 37–46. doi: 10.1177/001316446002000104
Cohen, J. (1988). Statistical Power Analysis for the Behavioral Sciences. Hillsdale: N.J. L. Erlbaum Associates.
Cueva, A. S., Galhardoni, R., Cury, R. G., Parravano, D. C., Correa, G., Araujo, H., et al. (2016). Normative data of cortical excitability measurements obtained by transcranial magnetic stimulation in healthy subjects. Neurophysiol. Clin. 46, 43–51. doi: 10.1016/j.neucli.2015.12.003
Di Lazzaro, V., Oliviero, A., Pilato, F., Saturno, E., Dileone, M., Mazzone, P., et al. (2004). The physiological basis of transcranial motor cortex stimulation in conscious humans. Clin. Neurophysiol. 115, 255–266. doi: 10.1016/j.clinph.2003.10.009
Di Lazzaro, V., Pellegrino, G., Di Pino, G., Corbetto, M., Ranieri, F., Brunelli, N., et al. (2015). Val66Met BDNF gene polymorphism influences human motor cortex plasticity in acute stroke. Brain Stimul. 8, 92–96. doi: 10.1016/j.brs.2014.08.006
Di Lorenzo, F., Bonn,ì, S., Picazio, S., Motta, C., Caltagirone, C., Martorana, A., et al. (2020). Effects of cerebellar theta burst stimulation on contralateral motor cortex excitability in patients with Alzheimer's disease. Brain Topogr. 33, 613–617. doi: 10.1007/s10548-020-00781-6
Egger, M., Davey Smith, G., Schneider, M., and Minder, C. (1997). Bias in meta-analysis detected by a simple, graphical test. BMJ 315, 629–634. doi: 10.1136/bmj.315.7109.629
Field, A. P., and Gillett, R. (2010). How to do a meta-analysis. Br. J. Math. Stat. Psychol. 63, 665–694. doi: 10.1348/000711010X502733
Fitzgerald, P. B., Brown, T. L., Daskalakis, Z. J., Chen, R., and Kulkarni, J. (2002). Intensity-dependent effects of 1 Hz rTMS on human corticospinal excitability. Clin. Neurophysiol. 113, 1136–1141. doi: 10.1016/S1388-2457(02)00145-1
Fitzgerald, P. B., Fountain, S., and Daskalakis, Z. J. (2006). A comprehensive review of the effects of rTMS on motor cortical excitability and inhibition. Clin. Neurophysiol. 117, 2584–2596. doi: 10.1016/j.clinph.2006.06.712
Fratello, F., Veniero, D., Curcio, G., Ferrara, M., Marzano, C., Moroni, F., et al. (2006). Modulation of corticospinal excitability by paired associative stimulation: reproducibility of effects and intraindividual reliability. Clin. Neurophysiol. 117, 2667–2674. doi: 10.1016/j.clinph.2006.07.315
Fried, P. J., Jannati, A., Davila-Pérez, P., and Pascual-Leone, A. (2017). Reproducibility of single-pulse, paired-pulse, and intermittent theta-burst TMS measures in healthy aging, type-2 diabetes, and Alzheimer's disease. Front. Aging Neurosci. 9, 263. doi: 10.3389/fnagi.2017.00263
Goldsworthy, M. R., Hordacre, B., Rothwell, J. C., and Ridding, M. C. (2021). Effects of rTMS on the brain: is there value in variability? Cortex 139, 43–59. doi: 10.1016/j.cortex.2021.02.024
Goldsworthy, M. R., Müller-Dahlhaus, F., Ridding, M. C., and Ziemann, U. (2014). Inter-subject variability of LTD-like plasticity in human motor cortex: a matter of preceding motor activation. Brain Stimul. 7, 864–870. doi: 10.1016/j.brs.2014.08.004
Guerra, A., López-Alonso, V., Cheeran, B., and Suppa, A. (2020). Variability in non-invasive brain stimulation studies: Reasons and results. Neurosci. Lett. 719, 133330. doi: 10.1016/j.neulet.2017.12.058
Hallett, M. (2000). Transcranial magnetic stimulation and the human brain. Nature 406, 147–150. doi: 10.1038/35018000
Hamada, M., Murase, N., Hasan, A., Balaratnam, M., and Rothwell, J. C. (2013). The role of interneuron networks in driving human motor cortical plasticity. Cereb Cortex 23, 1593–1605. doi: 10.1093/cercor/bhs147
Hermsen, A. M., Haag, A., Duddek, C., Balkenhol, K., Bugiel, H., Bauer, S., et al. (2016). Test-retest reliability of single and paired pulse transcranial magnetic stimulation parameters in healthy subjects. J. Neurol. Sci. 362, 209–216. doi: 10.1016/j.jns.2016.01.039
Herwig, U., Padberg, F., Unger, J., Spitzer, M., and Schönfeldt-Lecuona, C. (2001). Transcranial magnetic stimulation in therapy studies: examination of the reliability of “standard” coil positioning by neuronavigation. Biol. Psychiat. 50, 58–61. doi: 10.1016/S0006-3223(01)01153-2
Hinder, M. R., Goss, E. L., Fujiyama, H., Canty, A. J., Garry, M. I., Rodger, J., et al. (2014). Inter- and Intra-individual variability following intermittent theta burst stimulation: implications for rehabilitation and recovery. Brain Stimul. 7, 365–371. doi: 10.1016/j.brs.2014.01.004
Huang, Y.-Z., Edwards, M. J., Rounis, E., Bhatia, K. P., and Rothwell, J. C. (2005). Theta burst stimulation of the human motor cortex. Neuron 45, 201–206. doi: 10.1016/j.neuron.2004.12.033
Huang, Y.-Z., Rothwell, J. C., Edwards, M. J., and Chen, R.-S. (2008). Effect of physiological activity on an NMDA-dependent form of cortical plasticity in human. Cereb Cortex 18, 563–570. doi: 10.1093/cercor/bhm087
Hunter, J. E., and Schmidt, F. L. (2000). Fixed effects vs. random effects meta-analysis models: implications for cumulative research knowledge. Int. J. Select. Assess. 8, 275–292. doi: 10.1111/1468-2389.00156
Iezzi, E., Conte, A., Suppa, A., Agostino, R., Dinapoli, L., Scontrini, A., et al. (2008). Phasic voluntary movements reverse the aftereffects of subsequent theta-burst stimulation in humans. J. Neurophysiol. 100, 2070–2076. doi: 10.1152/jn.90521.2008
Jannati, A., Fried, P. J., Block, G., Oberman, L. M., Rotenberg, A., and Pascual-Leone, A. (2019). Test-retest reliability of the effects of continuous theta-burst stimulation. Front. Neurosci. 13, 447. doi: 10.3389/fnins.2019.00447
Julkunen, P., Säisänen, L., Danner, N., Niskanen, E., Hukkanen, T., and Mervaala, E. et al. (2009). Comparison of navigated and non-navigated transcranial magnetic stimulation for motor cortex mapping, motor threshold and motor evoked potentials. Neuroimage 44, 790–795. doi: 10.1016/j.neuroimage.2008.09.040
Koch, G., Esposito, Z., Kusayanagi, H., Monteleone, F., Codecá, C., Di Lorenzo, F., et al. (2011). CSF tau levels influence cortical plasticity in Alzheimer's disease patients. J. Alzheimers Dis. 26, 181–186. doi: 10.3233/JAD-2011-110116
Koo, T. K., and Li, M. Y. (2016). A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J. Chiropr. Med. 15, 155–163. doi: 10.1016/j.jcm.2016.02.012
Lang, N., Harms, J., Weyh, T., Lemon, R. N., Paulus, W., Rothwell, J. C., et al. (2006). Stimulus intensity and coil characteristics influence the efficacy of rTMS to suppress cortical excitability. Clin. Neurophysiol. 117, 2292–2301. doi: 10.1016/j.clinph.2006.05.030
Maeda, F., Keenan, J. P., Tormos, J. M., Topka, H., and Pascual-Leone, A. (2000). Modulation of corticospinal excitability by repetitive transcranial magnetic stimulation. Clin. Neurophysiol. 111, 800–805. doi: 10.1016/S1388-2457(99)00323-5
Modugno, N., Curr,à, A., Conte, A., Inghilleri, M., Fofi, L., Agostino, R., et al. (2003). Depressed intracortical inhibition after long trains of subthreshold repetitive magnetic stimuli at low frequency. Clin. Neurophysiol. 114, 2416–2422. doi: 10.1016/S1388-2457(03)00262-1
Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., et al. (2021). The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. Syst. Rev. 10, 89. doi: 10.1186/s13643-021-01626-4
Pascual-Leone, A., Valls-Sole, J., Wassermann, E. M., and Hallett, M. (1994). Responses to rapid-rate transcranial magnetic stimulation of the human motor cortex. Brain. 117, 847–58. doi: 10.1093/brain/117.4.847
Patel, R., Silla, F., Pierce, S., Theule, J., and Girard, T. A. (2020). Cognitive functioning before and after repetitive transcranial magnetic stimulation (rTMS): A quantitative meta-analysis in healthy adults. Neuropsychologia 141, 107395. doi: 10.1016/j.neuropsychologia.2020.107395
Peinemann, A., Reimer, B., Löer, C., Quartarone, A., Münchau, A., Conrad, B., et al. (2004). Long-lasting increase in corticospinal excitability after 1800 pulses of subthreshold 5 Hz repetitive TMS to the primary motor cortex. Clin. Neurophysiol. 115, 1519–1526. doi: 10.1016/j.clinph.2004.02.005
Pellegrini, M., Zoghi, M., and Jaberzadeh, S. (2018). The effect of transcranial magnetic stimulation test intensity on the amplitude, variability and reliability of motor evoked potentials. Brain Res. 1700, 190–198. doi: 10.1016/j.brainres.2018.09.002
Perellón-Alfonso, R., Kralik, M., Pileckyte, I., Princic, M., Bon, J., Matzhold, C., et al. (2018). Similar effect of intermittent theta burst and sham stimulation on corticospinal excitability: A 5-day repeated sessions study. Eur. J. Neurosci. 48, 1990–2000. doi: 10.1111/ejn.14077
Pitcher, J. B., Ogston, K. M., and Miles, T. S. (2003). Age and sex differences in human motor cortex input-output characteristics. J. Physiol. 546, 605–613. doi: 10.1113/jphysiol.2002.029454
Prei, K., Kanig, C., Osnabrügge, M., Langguth, B., Mack, W., Abdelnaim, M., et al. (2023). Limited evidence for validity and reliability of non-navigated low and high frequency rTMS over the motor cortex. medRxiv[Preprint]. doi: 10.1101/2023.01.24.23284951
Rosenthal, R. (1979). The file drawer problem and tolerance for null results. Psychol. Bull. 86, 638–641. doi: 10.1037/0033-2909.86.3.638
Rossini, P. M., Desiato, M. T., and Caramia, M. D. (1992). Age-related changes of motor evoked potentials in healthy humans: Non-invasive evaluation of central and peripheral motor tracts excitability and conductivity. Brain Res. 593, 14–19. doi: 10.1016/0006-8993(92)91256-E
Rossini, P. M., and Rossi, S. (1998). Clinical applications of motor evoked potentials. Electroencephalogr. Clin. Neurophysiol. 106, 180–194. doi: 10.1016/S0013-4694(97)00097-7
Sale, M. V., Ridding, M. C., and Nordstrom, M. A. (2007). Factors influencing the magnitude and reproducibility of corticomotor excitability changes induced by paired associative stimulation. Exp. Brain Res. 181, 615–626. doi: 10.1007/s00221-007-0960-x
Sale, M. V., Ridding, M. C., and Nordstrom, M. A. (2008). Cortisol inhibits neuroplasticity induction in human motor cortex. J. Neurosci. 28, 8285–8293. doi: 10.1523/JNEUROSCI.1963-08.2008
Schilberg, L., Schuhmann, T., and Sack, A. T. (2017). Interindividual variability and intraindividual reliability of intermittent theta burst stimulation-induced neuroplasticity mechanisms in the healthy brain. J Cogn. Neurosci. 29, 1022–1032. doi: 10.1162/jocn_a_01100
Schoisswohl, S., Langguth, B., Hebel, T., Abdelnaim, M. A., Volberg, G., and Schecklmann, M. (2021). Heading for personalized rTMS in tinnitus: reliability of individualized stimulation protocols in behavioral and electrophysiological responses. JPM. 11, 536. doi: 10.3390/jpm11060536
Shulga, A., Lioumis, P., Kirveskari, E., Savolainen, S., and Mäkelä, J. P. (2021). A novel paired associative stimulation protocol with a high-frequency peripheral component: A review on results in spinal cord injury rehabilitation. Eur. J. Neurosci. 53, 3242–3257. doi: 10.1111/ejn.15191
Siebner, H. R., and Rothwell, J. (2003). Transcranial magnetic stimulation: new insights into representational cortical plasticity. Exp. Brain Res. 148, 1–16. doi: 10.1007/s00221-002-1234-2
Sommer, M., Wu, T., Tergau, F., and Paulus, W. (2002). Intra- and interindividual variability of motor responses to repetitive transcranial magnetic stimulation. Clin. Neurophysiol. 113, 265–269. doi: 10.1016/S1388-2457(01)00726-X
Todd, G., Kimber, T. E., Ridding, M. C., and Semmler, J. G. (2010). Reduced motor cortex plasticity following inhibitory rTMS in older adults. Clin. Neurophysiol. 121, 441–447. doi: 10.1016/j.clinph.2009.11.089
Vallence, A.-M., Goldsworthy, M. R., Hodyl, N. A., Semmler, J. G., Pitcher, J. B., and Ridding, M. C. (2015). Inter- and intra-subject variability of motor cortex plasticity following continuous theta-burst stimulation. Neuroscience 304, 266–278. doi: 10.1016/j.neuroscience.2015.07.043
Vernet, M., Bashir, S., Yoo, W.-K., Oberman, L., Mizrahi, I., Ifert-Miller, F., et al. (2014). Reproducibility of the effects of theta burst stimulation on motor cortical plasticity in healthy participants. Clin. Neurophysiol. 125, 320–326. doi: 10.1016/j.clinph.2013.07.004
Keywords: neuromodulation, cortical excitability, rTMS, variability, motor evoked potentials, reliability, protocols
Citation: Kanig C, Osnabruegge M, Schwitzgebel F, Litschel K, Seiberl W, Mack W, Schoisswohl S and Schecklmann M (2023) Retest reliability of repetitive transcranial magnetic stimulation over the healthy human motor cortex: a systematic review and meta-analysis. Front. Hum. Neurosci. 17:1237713. doi: 10.3389/fnhum.2023.1237713
Received: 09 June 2023; Accepted: 08 August 2023;
Published: 13 September 2023.
Edited by:
Chris Baeken, Ghent University, BelgiumReviewed by:
Francesco Di Lorenzo, Santa Lucia Foundation (IRCCS), ItalyEva-Maria Pool, Alexianer Krefeld GmbH, Germany
Copyright © 2023 Kanig, Osnabruegge, Schwitzgebel, Litschel, Seiberl, Mack, Schoisswohl and Schecklmann. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Carolina Kanig, Y2Fyb2xpbmEua2FuaWcmI3gwMDA0MDt1bmlidy5kZQ==