- 1Department of Control and Computer Engineering, Polytechnic University of Turin, Turin, Italy
- 2Department of Human Neuroscience, Sapienza University of Rome, Rome, Italy
- 3IRCCS Neuromed Institute, Pozzilli, Italy
- 4Department of Electronic Engineering, University of Rome Tor Vergata, Rome, Italy
Introduction: The analysis of vocal samples from patients with Parkinson's disease (PDP) can be relevant in supporting early diagnosis and disease monitoring. Intriguingly, speech analysis embeds several complexities influenced by speaker characteristics (e.g., gender and language) and recording conditions (e.g., professional microphones or smartphones, supervised, or non-supervised data collection). Moreover, the set of vocal tasks performed, such as sustained phonation, reading text, or monologue, strongly affects the speech dimension investigated, the feature extracted, and, as a consequence, the performance of the overall algorithm.
Methods: We employed six datasets, including a cohort of 176 Healthy Control (HC) participants and 178 PDP from different nationalities (i.e., Italian, Spanish, Czech), recorded in variable scenarios through various devices (i.e., professional microphones and smartphones), and performing several speech exercises (i.e., vowel phonation, sentence repetition). Aiming to identify the effectiveness of different vocal tasks and the trustworthiness of features independent of external co-factors such as language, gender, and data collection modality, we performed several intra- and inter-corpora statistical analyses. In addition, we compared the performance of different feature selection and classification models to evaluate the most robust and performing pipeline.
Results: According to our results, the combined use of sustained phonation and sentence repetition should be preferred over a single exercise. As for the set of features, the Mel Frequency Cepstral Coefficients demonstrated to be among the most effective parameters in discriminating between HC and PDP, also in the presence of heterogeneous languages and acquisition techniques.
Conclusion: Even though preliminary, the results of this work can be exploited to define a speech protocol that can effectively capture vocal alterations while minimizing the effort required to the patient. Moreover, the statistical analysis identified a set of features minimally dependent on gender, language, and recording modalities. This discloses the feasibility of extensive cross-corpora tests to develop robust and reliable tools for disease monitoring and staging and PDP follow-up.
1. Introduction
Parkinson's disease (PD) is a neurodegenerative disorder caused by progressive and irreversible degeneration in the substantia nigra pars compacta region of the midbrain. For reasons not yet fully understood, the incidence of this disease has been rapidly increasing over the last two decades, with current incidence rates ranging from 31 to 328 cases per 100,000 inhabitants (1). Despite being idiopathic, PD is known to be prone to several risk factors, among which age and gender are relevant (2) According to recent evidence, PD chiefly occurs at the age of 57/60, with a risk of developing the disease that rises exponentially with ageing. Gender is also an established risk factor, with a male-female ratio of about 3:2 (3).
PDP usually face several motor and non-motor symptoms (4–6), including tremors at rest, bradykinesia, rigidity, dyskynesia (7), gait impairment (8), postural instability, sleep alterations, and voice disorders. Among these, it was estimated that vocal dysfunctions, usually grouped under the definition of dysarthria and dysphonia (9–11), affect ~90% of the PDP population. The incidence of these alterations, as well as the feasibility of rapid, non-invasive, and inexpensive signal collection, have made voice signal analysis a powerful and effective tool in supporting clinical practice (11, 12).
Indeed, voice production is performed through a complex sequence of movements that shape the excitation source to produce the perceived final sound. Any alteration in the control or execution of these movements turns into measurable alterations of the vocal signal, from which information about the patient's health status can be derived (13). In this context, a broad body of literature has been focused on the development of automatic tools for PDP vocal analysis, with promising results in disease staging, monitoring, and follow-up. Among these, the vast majority of the studies demonstrated the effectiveness of vocal-derived information as biomarkers for neurodegeneration, by employing mainly sustained phonation and continuous speech (14).
Sustained phonation is a simple and easily reproducible task that enhances the analysis of the glottal source vibration. Despite the absence of validated guidelines that clearly indicate which vowel has to be preferred, the vast majority of studies on PD vocal analysis employ the sustained phonation of the vowel /a/ (15). In this context, in Naranjo et al. (16) the authors proposed a model based on 44 acoustic features extracted from the sustained vowel phonation of 80 participants. These features were input to a Bayesian classifier, achieving 75.2% accuracy. Similarly, the studies conducted in (17) and (18) employed speech features from the freely available University California Irvine (UCI) dataset (188 PDP and 64 Healthy Controls—HC) (17) in conjunction with a Support Vector Machine (SVM) and a Logistic Regression Stacking Ensemble, respectively. The models yielded 84 and 92.22% classification accuracy.
Despite the excellent performance associated with the sustained phonation task, the lack of involvement of the articulators makes the exercise somehow simplistic, hence not suitable for a more complete analysis of speech production. In fact, some studies pointed out that reading sentences or monologues can help achieving a deeper insight, as they contain additional motion information and require more precise coordination (19). Indeed, a continuous speech analysis can improve the detection of changes in volume, pitch, and timing. In this context, several studies proved the effectiveness of running speech (20), reading text (21), or isolated word repetitions (22) with average classification accuracies between 80 and 95% (19). Nevertheless, while the sustained phonation task is usually considered to be minimally influenced by language, the information achievable from continuous speech embeds several complexities. Indeed, the same algorithm applied to different languages can provide different results. In this context, although many articles deploy PD speech analysis tools employing participants from various nationalities, only a few studies perform a direct language comparison. Among these, in (23) the authors investigated continuous speech samples from five different nationalities (i.e., Czech, English, German, French, and Italian) and identified changes in voice quality, articulation, and speed of speech. Recently, the authors in (24) carried out an analysis of five datasets in five different languages (i.e., Italian, Hebrew, English, Czech, and Spanish), encompassing 241 PDP and 265 HC. The study aimed not only to perform cross-language experiments, but also to verify the feasibility of a model trained on a heterogeneous dataset including participants from different nationalities and recorded in different conditions. This latter experiment yielded 75% classification accuracy provided by an Extreme Gradient Boosting (XGB) classifier. However, the performance is reported on a 10-fold cross-validation (CV), while no additional tests were performed on previously unseen samples.
In the present work, we aim at identifying the most adequate speech tasks as well as to investigate the most effective and trustworthy features that prove to be robust against external co-factors such as language, gender, and data collection modality. More in detail, we conducted a performance comparison between different types of speech tasks or their combination, to identify a speech protocol that can capture the vocal alterations of parkinsonian patients, while minimizing the effort required to the participants. Moreover, we employed heterogeneous corpora including participants from three different nationalities, and conducted an in-depth statistical analysis to assess the influence of participants and data-collection characteristics on the feature extracted and the model performance.
2. Materials and methods
2.1. Materials
We employed six different and heterogeneous corpora, including a total of 354 participants (176 PDP and 178 HC) from three different nationalities. For the sake of brevity, in this work we refer to each dataset as the language of the speakers it contains. In the case of multiple datasets under the same nationality, we made a distinction across the city where the corpus was recorded.
The Italian Bari corpus refers to the Italian Parkinson's Voice and Speech (IPVS), an open-access dataset that is available under the Creative Commons Attribution License (CC BY 4.0). Inclusion criteria and other information are reported in Dimauro et al. (25).
The Italian Torino1 and Italian Torino2 corpora refer to two private datasets recorded by ourselves. Participants were enrolled at A.O.U Città della Salute e della Scienza di Torino and Associazione Amici Parkinsoniani Piemonte Onlus. The inclusion criteria were: a clinical diagnosis of idiopathic PD with vocal signs and symptoms and no major cognitive impairment or other conditions preventing the patient from correctly accomplishing the task. To evaluate the efficiency of the proposed algorithms in the real-world scenario, both corpora were registered under suboptimal conditions. More in detail, Italian Torino1 includes samples collected in a non-supervised manner through a web application that guided the users in the execution of the same tasks encompassed in Italian Bari. As for Italian Torino2, audio samples were collected in a quiet room through an iPhone 12 under the supervision of an operator. Participants were instructed to sit in a relaxed position with the back and arms adhering to the back- and arm-rest (26) and perform the sustained phonation of the vowel /a/ at a comfortable volume with the smartphone placed ~10 cm to the mouth. The data collection of the two datasets were conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of the A.O.U Città della Salute e della Scienza di Torino (approval number 00384/2020). Participants received detailed information on the study purposes and execution. Informed consent for the observational study was obtained; demographic and clinical data were noted anonymously.
The Spanish corpus refers to the PC-GITA dataset, a private set of vocal recordings whose details are reported in Orozco-Arroyave et al. (27).
The Czech corpus is a publicly dataset that is available under the Creative Commons Attribution License (CC BY 4.0). Additional and more detailed information can be found in Hlavnicka et al. (28).
The Italian Rome is a private dataset detailed in Suppa et al. (29).
Table 1 summarizes the characteristics of the datasets, including numerosity, composition, and equipment employed during the data-collection procedure.
Data analysis was carried out in Python employing Praat (by Paul Boersma and David Weenink, Phonetic Sciences, University of Amsterdam) for pre-processing and feature extraction. The Parselmouth library was used as an interface to the Praat internal code.
2.2. Methods
2.2.1. Pre-processing and feature extraction
Data included in the six corpora were characterized by different sampling rates. Thereafter, all the recordings were firstly down-sampled to 16 kHz to assure consistency. The signal amplitudes were normalized in the range [0, 1] to prevent the speaker-microphone distance from affecting the model. We manually removed the initial and final silence regions; hence, no further preparatory steps were required. Finally, for the analysis of the phrases-repetition task, we employed the Praat software to detect start- and end-points of voiced regions. After identifying and merging voiced regions, we framed each signal into 25 ms windows with 10 ms overlap and extracted features from each segment (30–32). This set encompassed, among the others, periodicity measures (i.e., fundamental frequency F0, first three formants, and their bandwidths) and noise measures including Harmonc to Noise Ratio (HNR), Cepstral Peak Prominence (CPP), and Glottal to Noise Excitation ratio (GNE). Moreover, we extracted spectral and cepstral features together with their first and second derivative, which often proved to be effective in PD-related speech analysis (15). Finally, we grouped the features into one feature vector and calculated five statistics (i.e., mean value, median value, standard deviation, kurtosis, and skewness). It is worth noting that jitter and shimmer variants were evaluated from the entire signal, given that their definition already embeds a comparison among contiguous frames. Table 2 reports an overview of the employed features together with their description. Min-max normalization was applied to the whole set of features to scale them to the same range.
2.2.2. Experiment 1: identification of the speech protocol
For the first aim of this work, i.e., to identify a set of speech tasks that can capture PDP vocal alterations while minimizing the effort required to the patients, we employed the four corpora that included at least two speech exercises (i.e., Italian Bari, Italian Torino1, Spanish, and Italian Roma). This analysis was carried out in two subsequent steps. First, we assessed whether different variants of the same speech tasks (e.g., different vowels for the sustained vowel phonation, different phrases for the sentence repetition) exhibit different effectiveness in discriminating PDP and HC. If so, we wanted to verify if the same evidence occurs in multiple corpora, regardless of the speakers' language or data-collection modality. Thereafter, we investigated whether different speech tasks or their combination can provide better performance in automatically distinguishing PDP from HC. Due to the composition of the included corpora, we used the Italian Bari and Spanish datasets to compare /a/ and /e/ vowels. Similarly, we employed Italian Bari, Italian Torino1, Spanish, and Italian Roma to identify possible difference within a set of phonetically balanced phrases. It is worth noting that the first two datasets included the pronunciation of the same set of sentences, hence, after performing the comparison between the two vowels, we decided to merge them into a single dataset. We remark that, due to the differences in the collection procedure, a stratified fusion was applied to maintain the same proportion of the two datasets in any subsequent split. The best vowels and phrases resulting from this analysis underwent a further comparison to assess whether a single task or their combination can better distinguish between the two classes. The combination of the two tasks is obtained through an early fusion of the related features. Indeed, in a previous work (22) we demonstrated that this approach is more effective than a late fusion through majority voting. Also in this case, the results obtained from different corpora were taken into account to test the robustness of the obtained evidence. The dataset Italian Roma only contains the phonation of a single vowel (/e/), hence this latter was assumed to be the most significant. Due to the balanced distribution of the corpora used in this study, we compared the effectiveness of each speech task as the accuracy obtained from a binary classification model trained with features extracted from the task at hand. The pipeline employed included a feature selection step, performed through the Boruta algorithm, and a classification step.
The Boruta algorithm is a wrapper technique that iteratively uses the classification scores of a Random Forest (RF) model to identify the optimal set of features for the application at hand (33). For each input variable, the algorithm creates a shadow attribute by shuffling the values of the original features across objects. The classification is then performed employing the extended set of features. The chief assumption under the algorithm is that adding randomness to the system and analyzing its impact on the model can highlight the most significant features (34).
As for the automatic binary classification between PDP and HC, we tested ten different algorithms: k-Nearest Neighbors (KNN), SVM, Gaussian Process (GP), Decision Tree (DT), RF, Artificial Neural Network (ANN), Naive Bayes (NB), Linear Discriminant Analysis (LDA), AdaBoost (ADA), and XGB. To minimize the influence of the classification model on the comparison between speech tasks, we decided to average their performance and use the returned value for comparisons. To avoid the weak generalization capability of the model, we performed feature selection and model training on 70% of the original dataset and evaluated the performance through a 10-fold cross-validation (CV). The remaining 30% of participants were employed only during the testing phase, so the speaker independence was guaranteed (i.e., recordings from each participant were either in train or in test), without further optimization.
The speech task yielding the best performance was assessed through the accuracy metric. F1-score, Sensibility, Specificity, and Area Under the Curve (AUC) were also computed for comparison purposes.
2.2.3. Experiment 2: analysis of the features effectiveness and robustness
In the second experiment, we performed a series of statistical tests to evaluate the feature effectiveness in discriminating between PDP and HC and the robustness against external factors. We first applied the U-Mann–Whitney test to Italian Bari, Italian Torino1, Spanish, Italian Torino2, and Czech corpora and identified the features that exhibited a statistically significant difference (p < 0.05) between PDP and HC in at least three datasets. Thereafter, we created a unified dataset and repeated the test to study the influence of the dataset heterogeneity on the selected features.
Finally, we performed a Kruskal-Wallis test to assess the presence of statistically significant differences (p < 0.05) in the feature distribution due to external factors such as gender, language, and the data-collection modality. In order to avoid possible bias due to differences between HC and PDP, we performed the test separately on the two subgroups and considered a feature as exhibiting significant differences only if the null hypothesis was rejected in both populations. Based on the results of the statistical analyses, a resulting set of effective and robust features underwent feature selection and classification steps. The same pipeline employed in Experiment 1 was applied. It is worth noting that only speech samples associated with the sustained vowel phonation were employed for this experiment to minimize possible bias due to language-dependent phonemes. Moreover, also in this case, to avoid the weak generalization capability of the model, we trained the model on 70% of the original dataset and evaluated the performance using a 10-fold CV. The remaining 30% of participants was employed only during the testing phase, without any further optimization. In Figure 1, we report a schematic overview of the two carried out experiments.
3. Results
3.1. Experiment 1
In Figure 2, we report the results of the first comparison performed in Experiment 1. In particular, in Figures 2A, B the classification accuracy for vowel and phrase comparisons are depicted. In Figure 3, we report the results of the comparison among the best-performing speech exercises and their combination. After selecting the best speech protocol, we performed an additional comparison in order to identify the classification model that yields the best performance. To this aim, for each model employed, we computed the average classification accuracy across the three tested corpora. According to our analysis, the best models resulted to be KNN and GP, which lead to 91% average accuracy on validation sets. The two best-performing classifiers were then employed to compute a comprehensive set of metrics on validation and tests set, in order to evaluate the capability of the model to predict new and previously unseen samples (Table 3).
Figure 2. Experiment 1: results of the comparison across different vowels (A) and different phrases (B) in differentiating between HC and PDP.
Figure 3. Experiment 1: results of the comparison across different speech tasks or their combination.
3.2. Experiment 2
In Table 4, we report the results of the performed statistical tests. For the sake of brevity, in the case of the unified dataset we include only those features that exhibited a statistically significant inter-class difference and proved robust with respect to language, gender, and dataset characteristics. Table 5 reports the classification accuracy of the two best-performing models for two different feature sets. The first one is obtained from the application of the Boruta algorithm to the set of features extracted in the previous analysis (i.e., those selected in at least three corpora or in the unified dataset and minimally influenced by external factors). The second feature set is obtained by adding external co-variables (i.e., gender, language, data collection modality) prior to the feature selection procedure. In fact, it has been previously demonstrated that the addition of such co-factors can improve the generalization capability of the feature selection procedure (15). For each feature sets obtained from the application of the Boruta algorithm, we performed a post-hoc analysis through the computation of the Spearman correlation coefficient, in order to investigate their importance in the application at hand. According to our findings, the standard deviation of the first formant showed the highest correlation with class (r:0.40, P: ≤ 0.001), followed by mean value of the spectral center of gravity (r:−0.35, P: ≤ 0.001) and the local Jitter (r:0.27, P: ≤ 0.001).
4. Discussion
Our study confirmed the feasibility of an automatic classification between HC and PDP, identified a series of speech exercises effective in capturing vocal alterations in PDP, and tested the robustness of acoustic features against external co-factors such as language, gender, and data-collection modality.
According to our findings, an adequate speech protocol should include the phonation of a sustained vowel and the pronunciation of a phonetically balanced phrase. As shown in Figure 3, this combination yields a performance improvement up to 13.6% compared to the single vowel phonation and up to 5.08% compared to the pronunciation of a single phrase. Due to the high complexity of the articulatory dimension of the speech process, we performed an additional comparison embedding the second best-performing phonetically balanced sentence. However, no performance improvement was obtained in any of the employed datasets. As for the specific type of speech tasks, the vowel /a/ exhibited better discriminatory capability between PDP and HC than the vowel /e/Figure 2A. Despite the same trend can be observed in the Italian Bari and the Spanish datasets, this latter achieved lower classification performance, probably due to the different dataset composition. Moreover, despite the the fact that the same algorithm was employed for all the datasets included in the experiment, different performance was registered across different corpora. Indeed, from a comparison of Figure 2A and Figure 3, we can observe a non-negligible performance impairment associated with the single vowel phonation when adding the Italian Torino1 to the Italian Bari corpus. More in detail, in Figure 2A we report the performance obtained on the Italian Bari dataset both alone and in conjuction with the Italian Torino 1 corpus. According to our findings, despite a careful stratification procedure was employed, the inclusion of samples from the latter corpus led to an impairment of the classification accuracy. Despite the need of larger datasets to obtain statistically relevant results, this evidence points out the impact of the data collection modality. In fact, as previously demonstrated in (35) and in (36), despite a careful stratified merging procedure, the usage of vocal recordings collected in different conditions impacts the classification performance.
As for the comparison between different phonetically balanced phrases, our results identify several sentences yielding optimal performance (see Figure 2B). However, except for the Italian Roma corpus that includes only three sentences, more than 10% improvement can be achieved by moving from the less- to the most-effective sentence. As for the Italian Bari and the Italian Torino1 corpora, the best performance is associated to phrase 3 (“Patrizia ha pranzato a casa to Fabio”—“Patrizia had dinner at Fabio's house”), while the worst performance is associated to phrase 8 (“Marco non è partito”—“Marco did not leave”). The comparison between these two sentences put into evidence the need for a more articulated task to effectively capture vocal alterations, as well as the need of embedding both occlusive and fricative sounds, as previously demonstrated in (36) and (37). Similarly, the comparison between phrase 1 (“Omar, que vive cerca, trajo miel”—“Omar, who lives near, brought honey”) and phrase 3 (“Los libros nuevos no caben en la mesa de la oficina”—“The new books do not fit in the office's table”) in the Spanish datasets, highlights that better performance is achieved using a sentence that includes more complex and articulated sounds, as well occlusive consonants.
The results of the classification performance yielded by the combination of best-performing vowel and sentence are reported in Table 3. KNN and GP proved to be the best classification models, with classification accuracy on a 10-fold CV ranging from 88.7 to 94.5%. No significant performance impairment is observed when moving to the test set, suggesting an optimal generalization capability. Moreover, the model exhibited similar performance on three different corpora, showing a good robustness. As for the selected features, the Boruta algorithm returned different sets of features for the three corpora, with a high prevalence of phrase-derived features. The comparison between the three subgroups suggests the effectiveness of spectral center of gravity, MFCC, and PLP coefficients.
As for the second experiment conducted in this study, in Tables 4, 5 we reported the results of the feature statistical analysis and the performed classification step, respectively. MFCC proved to be effective in discriminating between PDP and HC also in presence of heterogeneous datasets and did not show a statistically significant correlation with language, gender, or dataset taken into consideration. Similarly, despite F0 did not prove robust against external factors, the related differential amplitude measures (represented by Jitter features) is crucial in discriminating between the two classes and little influenced by language and dataset type. As for the classification step, XGB and RF outperformed the other considered classification models. A generalization capability improvement is associated to the use of feature set 2, which includes external co-variable such as language and gender before feature selection. Indeed, as previously reported (15), we assume that adding such co-factors before feature selection can provide pivotal clues during the procedure, thus leading to a final set of features that are more representative of the actual vocal alterations despite the peculiarities of the population considered. Also in this case, no significant performance impairment is observed when moving from the validation to the test set.
As expected, a performance impairment can be observed in Experiment 2 with respect to Experiment 1 (average classification accuracy on the validation sets of 77.3 and 70.2% in the two Experiments). Despite this impairment, which can be mainly ascribable to the heterogeneity of the dataset, the result confirms the feasibility of a classification algorithm trained on a unified dataset. Indeed, it is well known that one of the main problems encountered in the design of automatic tools for vocal pathology assessment concerns the database size, which can lead to feature selection and classification results excessively fitted on the population at hand (38). Moreover, although the use of highly homogeneous corpora would achieve better results, the same conditions can be hardly repeatable, thus limiting the actual usefulness in real-world scenarios. In this context, the statistical tests performed in Experiment 2, despite not exhaustive, provide pivotal clues on the influence of external co-factors on the acoustic features. This can help identifying those aspects against which stratification is necessary, and those aspects that may be managed deploying a proper algorithm (e.g., through a co-variable introduction before feature selection).
Despite the lack of similar studies in the related literature, we compared our classification results with those in (24). To the best of our knowledge, this is the only similar study conducted up to date. In this work, the authors achieved 75% classification accuracy on a 10-fold CV performed on a heterogeneous dataset including 241 PDP and 265 HC. However, additional analyses on a separate test set were not performed.
5. Limitations
Despite the promising results, we acknowledge the presence of several limitations that still must be addressed. Among these, we must mention the numerosity of the included corpora. In fact, despite the cardinality of the datasets used for the analysis allows to achieve statistically significant results and is generally higher than in similar studies, we plan to further increase the sample and test our findings on wider corpora. Moreover, due to the absence of a single protocol performed by speakers of different nationalities and collected in different conditions, we had to perform the second Experiment only on recording from sustained phonations, despite the evidence deriving from Experiment 1 suggests that the contribution of the repetition of a phonetically balanced phrase is crucial. Finally, in both Experiments, we performed binary comparisons and classification, while not addressing any disease staging task.
6. Conclusions and future works
In this work, we presented a comprehensive analysis to identify an adequate set of speech exercises that can effectively describe vocal alterations in patients with PD. The results on achieved on different corpora, including participants from different nationalities, showed that the joint use of the sustained phonation and the repetition of a phonetically balanced phrase should be preferred to the use of a single task. As for the specific tasks, the usage of the vowel /a/ together with an articulated sentence that embeds both occlusive and fricative sounds resulted effective, with 91% average classification accuracy on the three datasets. A post-hoc analysis of the features selected and input into the ML models revealed different sets for the three corpora, with a high prevalence of phrase-derived features. The comparison between the three subgroups indicated the effectiveness of spectral center of gravity, MFCC, and PLP coefficients.
The second aim of this study was to investigate the feasibility of a classification between HC and PDP employing a heterogeneous dataset and derive information about the feature trustworthiness and robustness against the speaker characteristics and the data-collection modalities. To this aim, we employed 4 corpora including participants from three different nationalities. According to our evidence, MFCC coefficients proved to be effective in discriminating between PDP and HC and did not show a statistically significant correlation with gender, language, and dataset type taken into consideration. Similarly, differential frequency measures represented by Jitter features proved to be crucial in discriminating between the two classes while being influenced only by gender of the speaker. As for the classification step, including external co-variables (such as language and gender before feature selection) improved the generalization capability of the model, yielding 70% accuracy both on train and test sets. Despite a 7% reduction with respect to the single datasets, these results disclose the option for a model trained on unified corpora. Indeed, one of the main limitations of the automatic tools for supporting clinical practice concerns data availability. The evidence from this work can be used to derive information about the external factors that have a stronger influence on the features extracted and the model performance, thus performing a specific stratification and improving the performance of the overall algorithm. In this context, future works will address the collection of additional datasets to increase the overall cardinality and test whether building gender- or data-collection-specific models can improve the classification performance. Moreover, we only performed the multi-lingual analysis on the vowel /a/ in order to control the sources of system complexity (i.e., difference encountered in diverse language may be due to the phoneme-specific characteristics), however, future studies will include also text-based analysis which, according to the results obtained on the single datasets, we expect to provide a performance improvement. Indeed, as previously reported in (39), although a broad body of literature pointed out the effectiveness of an analysis based on phonatory aspects of speech production, more recent studies highlighted the importance of including articulatory aspects as well, in order to allow a better understanding of the impairment. In fact, if on one hand the more explicit hallmarks of PD vocal alterations can be detected through a simple sustained vowel phonation, more complex tasks can reveal hidden clues of the vocal dimensions, such as the difficulty in performing finer movements than can be measured only in presence of a voiced/unvoiced traits alternations (36). At the same time, more complex tasks instrinsically embed several complexities that still have to be addressed, such as the difference in phonemes pronunciation which characterize different languages. In this context, future work will be in the direction of defining a speech protocol that, while including relevant phonetically balanced sounds which can inspect PDP vocal alterations, can be performed also by participants from different nationalities and cultural backgrounds. Finally, in both Experiments, we performed binary comparisons and classifications. Aware of the importance of the disease staging task, future work will also introduce information about the patient's health status and generalize the current evidence in the presence of multi-class analyses.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found here: the Italian Bari dataset used in this study can be found at: https://ieee-dataport.org/open-access/italian-parkinsons-voice-and-speech; the Czech dataset used in this study can be found at: https://figshare.com/articles/dataset/Synthetic_vowels_of_speakers_with_Parkinson_s_disease_and_Parkinsonism/7628819?file=14172821 vowels of speakers with Parkinson's disease and Parkinsonism/7628819.
Ethics statement
The studies involving human participants were reviewed and approved by Ethics Committee of the A.O.U Città della Salute e della Scienza di Torino (approval number 00384/2020). The patients/participants provided their written informed consent to participate in this study.
Author contributions
FAm and GO: conceptualization. SS and FAm: methodology, writing-original draft, software, formal analysis, investigation, writing-review, and editing. SS: resources. GO, AS, GS, GC, and FAs: data curation. FAm, GO, GC, and GS: supervision and project administration. All authors have read and agreed to the published version of the manuscript.
Acknowledgments
The authors would like to thank the Associazione Amici Parkinsoniani Piemonte Onlus, the Centro Regionale Esperto per la Malattia di Parkinson, Dipartimento di Neuroscienze e Salute Mentale A.O.U Citta della Salute e della Scienza di Torino for their contribution to the data collection procedure.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Bloem BR, Okun MS, Klein C. Parkinson's disease. Lancet. (2021) 397:2284–303. doi: 10.1016/S0140-6736(21)00218-X
2. Asci F, Costantini G, Di Leo P, Zampogna A, Ruoppolo G, Berardelli A, et al. Machine-learning analysis of voice samples recorded through smartphones: the combined effect of ageing and gender. Sensors. (2020) 20:5022. doi: 10.3390/s20185022
3. Massano J, Bhatia KP. Clinical approach to Parkinson's disease: features, diagnosis, and principles of management. Cold Spring Harb Perspect Med. (2012) 2:a008870. doi: 10.1101/cshperspect.a008870
4. Lazzaro GD, Ricci M, Al-Wardat M, Schirinzi T, Scalise S, Giannini F, et al. Technology-based objective measures detect subclinical axial signs in untreated, de novo Parkinson's disease. J Parkinson's Dis. (2020) 10:113–22. doi: 10.3233/JPD-191758
5. Ricci M, Lazzaro GD, Pisani A, Mercuri NB, Giannini F, Saggio G. Assessment of motor impairments in early untreated Parkinson's disease patients: the wearable electronics impact. IEEE J Biomed Health Informatics. (2020) 24:120–30. doi: 10.1109/JBHI.2019.2903627
6. Lazzaro GD, Ricci M, Saggio G, Costantini G, Schirinzi T, Alwardat M, et al. Technology-based therapy-response and prognostic biomarkers in a prospective study of a de novo Parkinson's disease cohort. NPJ Parkinson's Dis. (2021) 12:7. doi: 10.1038/s41531-021-00227-1
7. Ricci M, Giannini F, Saggio G, Cenci C, Lazzaro GD, Pisani A. A novel analytical approach to assess dyskinesia in patients with Parkinson disease. In: IEEE International Symposium on Medical Measurements and Applications, MeMeA (2018) - Proceedings. Rome (2018). doi: 10.1109/MeMeA.2018.8438666
8. Pietrosanti L, Calado A, Verrelli CM, Pisani A, Suppa A, Fattapposta F, et al. harmonic distortion aspects in upper limb swings during gait in Parkinson's disease. Electronics. (2023) 12:625. doi: 10.3390/electronics12030625
9. Costantini G, Leo PD, Asci F, Zarezadeh Z, Marsili L, Errico V, et al. Machine learning based voice analysis in spasmodic dysphonia: an investigation of most relevant features from specific vocal tasks. In: 14th International Conference on Bio-inspired Systems and Signal Processing. Vienna (2021). p. 103–13. doi: 10.5220/0010344601030113
10. Suppa A, Asci F, Saggio G, Marsili L, Casali D, Zarezadeh Z, et al. Voice analysis in adductor spasmodic dysphonia: objective diagnosis and response to botulinum toxin. Parkinsonism Relat Disord. (2020) 73:23–30. doi: 10.1016/j.parkreldis.2020.03.012
11. Fayad R, Hajj-Hassan M, Constantini G, Zarazadeh Z, Errico V, Pisani A, et al. Vocal Test Analysis for Assessing Parkinson's Disease at Early Stage. In: 2021 Sixth International Conference on Advances in Biomedical Engineering (ICABME). Werdanyeh: Institute of Electrical and Electronics Engineers Inc. (2021). p. 171–4. doi: 10.1109/ICABME53305.2021.9604891
12. Costantini G, Cesarini V, Leo PD, Amato F, Suppa A, Asci F, et al. Artificial intelligence-based voice assessment of patients with Parkinson's disease off and on treatment: machine vs. deep-learning comparison. Sensors. (2023) 23:2293. doi: 10.3390/s23042293
13. Saggio G, Costantini G. Worldwide healthy adult voice baseline parameters: a comprehensive review. J Voice. (2022) 36:637–49. doi: 10.1016/j.jvoice.2020.08.028
14. Ngo T, Pathirana PN, Horne MK, Power L, Szmulewicz DJ, Milne SC, et al. Balance deficits due to cerebellar ataxia: a machine learning and cloud-based approach. IEEE Trans Biomed Eng. (2021) 68:1507–17. doi: 10.1109/TBME.2020.3030077
15. Amato F, Saggio G, Cesarini V, Olmo G, Costantini G. Machine learning- and statistical-based voice analysis of Parkinson's disease patients: a survey. Expert Syst Appl. (2023) 219:119651–67. doi: 10.1016/j.eswa.2023.119651
16. Naranjo L, Pérez CJ, Campos-Roca Y, Martín J. Addressing voice recording replications for Parkinson's disease detection. Expert Syst Appl. (2016) 46:286–92. doi: 10.1016/j.eswa.2015.10.034
17. Sakar CO, Serbes G, Gunduz A, Tunc HC, Nizam H, Sakar BE, et al. A comparative analysis of speech signal processing algorithms for Parkinson's disease classification and the use of the tunable Q-factor wavelet transform. Appl Soft Comput J. (2019) 74:255–63. doi: 10.1016/j.asoc.2018.10.022
18. Mohammadi AG, Mehralian P, Naseri A, Sajedi H. Parkinson's disease diagnosis: the effect of autoencoders on extracting features from vocal characteristics. Array. (2021) 11:100079. doi: 10.1016/j.array.2021.100079
19. Moro-Velazquez L, Gomez-Garcia J, Dehak N, Godino-Llorente JI. New tools for the differential evaluation of Parkinson's disease using voice and speech processing. In: Proceedings of IberSPEECH 2021. Valladolid: IberSPEECH (2021). p. 165–9. doi: 10.21437/IberSPEECH.2021-36
20. Benba A, Jilbab A, Hammouch A. Voice analysis for detecting patients with Parkinson's disease using the hybridization of the best acoustic features. Int J Electr Eng Inform. (2016) 8:108–16.
21. Toye AA, Kompalli S. Comparative Study of Speech Analysis Methods to Predict Parkinson's Disease (2021). Available online at: http://arxiv.org/abs/2111.10207
22. Amato F, Borzì L, Olmo G, Rafael J, Arroyave O. An algorithm for Parkinson's disease speech classification based on isolated words analysis. Health Inf Sci Syst. (2021) 9:32. doi: 10.1007/s13755-021-00162-8
23. Rusz J, Hlavnička J, Novotný M, Tykalová T, Pelletier A, Montplaisir J, et al. Speech biomarkers in rapid eye movement sleep behavior disorder and Parkinson disease. Ann Neurol. (2021) 90:62–75. doi: 10.1002/ana.26085
24. Kovac D, Mekyska J, Aharonson V, Harar P, Galaz Z, Rapcsak S. Exploring Language-independent Digital Speech Biomarkers of Hypokinetic Dysarthria (2022). Available online at: https://doi.org/10.1101/2022.10.24.22281459
25. Dimauro G, Caivano D, Bevilacqua V, Girardi F, Napoletano V. VoxTester, software for digital evaluation of speech changes in Parkinson disease. In: IEEE International Symposium on Medical Measurements and Applications, MeMeA (2016) Proceedings. Benevento (2016). doi: 10.1109/MeMeA.2016.7533761
26. Saggio G, Sbernini L. New scenarios in human trunk posture measurements for clinical applications. In: 2011 IEEE International Symposium on Medical Measurements and Applications. Bari (2011). p. 13–7. doi: 10.1109/MeMeA.2011.5966651
27. Orozco-Arroyave JR, Arias-Londoño JD, Vargas-Bonilla JF, González-Rátiva MC, Nöth E. New Spanish speech corpus database for the analysis of people suffering from Parkinson's disease. In: Proceedings of the 9th International Conference on Language Resources and Evaluation, LREC 2014. Reykjavik (2014). p. 342–7.
28. Hlavnicka J, Cmejla R, Klempir J, Ruzicka E, Rusz J. Acoustic tracking of pitch, modal, and subharmonic vibrations of vocal folds in Parkinson's disease and parkinsonism. IEEE Access. (2019) 7:150339–54. doi: 10.1109/ACCESS.2019.2945874
29. Suppa A, Costantini G, Asci F, Leo PD, Al-Wardat MS, Lazzaro GD, et al. Voice in Parkinson's disease: a machine learning study. Front Neurol. (2022) 13:831428. doi: 10.3389/fneur.2022.831428
30. Khaskhoussy R, Ayed YB. Speech processing for early Parkinson's disease diagnosis: machine learning and deep learning-based approach. Soc Netw Anal Mining. (2022) 12:73. doi: 10.1007/s13278-022-00905-9
31. Tougui I, Jilbab A, El Mhamdi J. Machine learning smart system for Parkinson disease classification using the voice as a biomarker. Healthc Inf Res. (2022) 28:210–21. doi: 10.4258/hir.2022.28.3.210
32. Viswanathan R, Khojasteh P, Aliahmad B, Arjunan SP, Ragnav S, Kempster P, et al. Efficiency of voice features based on consonant for detection of Parkinson's Disease. In: 2018 IEEE Life Sciences Conference (LSC). Montreal, QC: IEEE (2018). p. 49–52. doi: 10.1109/LSC.2018.8572266
33. Tunc HC, Sakar CO, Apaydin H, Serbes G, Gunduz A, Tutuncu M, et al. Estimation of Parkinson's disease severity using speech features and extreme gradient boosting. Med Biol Eng Comput. (2020) 58:2757–73. doi: 10.1007/s11517-020-02250-5
34. Kursa MB, Rudnicki WR. Feature selection with the boruta package. J Stat Softw. (2010) 36:1–13. doi: 10.18637/jss.v036.i11
35. Carrón J, Campos-Roca Y, Madruga M, Pérez CJ. A mobile-assisted voice condition analysis system for Parkinson's disease: assessment of usability conditions. BioMed Eng Online. (2021) 20:114. doi: 10.1186/s12938-021-00951-y
36. Amato F, Borzi L, Olmo G, Artusi CA, Imbalzano G, Lopiano L. Speech impairment in Parkinson's disease: acoustic analysis of unvoiced consonants in Italian native speakers. IEEE Access. (2021) 9:1. doi: 10.1109/ACCESS.2021.3135626
37. Moro-Velazquez L, Gomez-Garcia JA, Godino-Llorente JI, Grandas-Perez F, Shattuck-Hufnagel S, Yagüe-Jimenez V, et al. Phonetic relevance and phonemic grouping of speech in the automatic detection of Parkinson's Disease. Sci Rep. (2019) 9:1–16. doi: 10.1038/s41598-019-55271-y
38. Gómez-Vilda P, Gómez-Rodellar A, Palacios-Alonso D, Rodellar-Biarge V, Álvarez Marquina A. The role of data analytics in the assessment of pathological speech: a critical appraisal. Appl Sci. (2022) 12:11095. doi: 10.3390/app122111095
Keywords: speech analysis, acoustic features, Parkinson's disease, machine learning, statistical analysis
Citation: Scimeca S, Amato F, Olmo G, Asci F, Suppa A, Costantini G and Saggio G (2023) Robust and language-independent acoustic features in Parkinson's disease. Front. Neurol. 14:1198058. doi: 10.3389/fneur.2023.1198058
Received: 31 March 2023; Accepted: 26 May 2023;
Published: 13 June 2023.
Edited by:
Giovanni Rizzo, IRCCS Institute of Neurological Sciences of Bologna (ISNB), ItalyReviewed by:
Gemma Moya Gale, Long Island University-Brooklyn, United StatesDecho Surangsrirat, National Science and Technology Development Agency (NSTDA), Thailand
Copyright © 2023 Scimeca, Amato, Olmo, Asci, Suppa, Costantini and Saggio. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Federica Amato, ZmVkZXJpY2EuYW1hdG8mI3gwMDA0MDtwb2xpdG8uaXQ=