- Department of Diagnostic and Interventional Radiology, University Hospital Würzburg, Würzburg, Germany
Purpose: Machine learning based on radiomics features has seen huge success in a variety of clinical applications. However, the need for standardization and reproducibility has been increasingly recognized as a necessary step for future clinical translation. We developed a novel, intuitive open-source framework to facilitate all data analysis steps of a radiomics workflow in an easy and reproducible manner and evaluated it by reproducing classification results in eight available open-source datasets from different clinical entities.
Methods: The framework performs image preprocessing, feature extraction, feature selection, modeling, and model evaluation, and can automatically choose the optimal parameters for a given task. All analysis steps can be reproduced with a web application, which offers an interactive user interface and does not require programming skills. We evaluated our method in seven different clinical applications using eight public datasets: six datasets from the recently published WORC database, and two prostate MRI datasets—Prostate MRI and Ultrasound With Pathology and Coordinates of Tracked Biopsy (Prostate-UCLA) and PROSTATEx.
Results: In the analyzed datasets, AutoRadiomics successfully created and optimized models using radiomics features. For WORC datasets, we achieved AUCs ranging from 0.56 for lung melanoma metastases detection to 0.93 for liposarcoma detection and thereby managed to replicate the previously reported results. No significant overfitting between training and test sets was observed. For the prostate cancer detection task, results were better in the PROSTATEx dataset (AUC = 0.73 for prostate and 0.72 for lesion mask) than in the Prostate-UCLA dataset (AUC 0.61 for prostate and 0.65 for lesion mask), with external validation results varying from AUC = 0.51 to AUC = 0.77.
Conclusion: AutoRadiomics is a robust tool for radiomic studies, which can be used as a comprehensive solution, one of the analysis steps, or an exploratory tool. Its wide applicability was confirmed by the results obtained in the diverse analyzed datasets. The framework, as well as code for this analysis, are publicly available under https://github.com/pwoznicki/AutoRadiomics.
Introduction
Over the past decades, the search for novel, quantitative imaging biomarkers has been an emerging topic in the research landscape, with the ultimate goal of leveraging the full potential of medical imaging and enabling more personalized medical care (1, 2). Within this field, radiomics has been identified as a potential way to mathematically extract clinically meaningful quantitative imaging biomarkers (so-called features) from medical images of different modalities (3–5). Combined with machine learning (ML), radiomics classifiers have been shown to accurately predict the diagnosis (6), prognosis (7), mutational status / genetic subtypes (8–10), histopathology (8), surgery (11), or treatment response (12). Consequently, there is a huge interest in the clinical and research field to translate the diagnostic and prognostic potential of radiomics to clinical patient care.
This interest has resulted in a large number of scientific publications being issued with a similarly large variety of methods and radiomics pipelines. Besides the inherent issue of model overfitting, which comes with any ML and big data application where the number of features usually considerably exceeds the number of samples in the training set, most radiomics studies also have been proven difficult to reproduce and validate. This may be also due to the large variety of methodology and the lack of an open-science mindset within the research community, with the datasets and code rarely published alongside the results.
Fortunately, an evolving body of open-science frameworks has been accumulating in recent years, and new initiatives aiming at standardization and reproducibility of different aspects of radiomics analysis and ML have been founded. For example, the Image Biomarker Standardization Initiative (IBSI) (3) has addressed the standardization of the radiomic feature extraction process, while the Workflow for Optimal Radiomics Classification (WORC) (2) has been developed in order to automate and standardize a typical radiomics (and ML) workflow.
Performing reproducible radiomics studies usually requires programming skills, since the most prevalent tools in the research community are written in Python language (1–3). This makes it very difficult for clinicians (who will be the ones responsible for clinical translation of trained models and classifiers) to perform radiomics studies by themselves or to simply “play around” with the data.
The aim of this study was to present an intuitive, open-source framework with an interactive user interface for reproducible radiomics workflow. We evaluated its performance on eight publicly available datasets covering varying clinical applications to prove that the framework is able to reproduce previously published studies. AutoRadiomics provides tools for every step of the radiomics workflow (including image segmentation, image processing, feature extraction, classification, and evaluation) with the ability to adjust each step of the workflow as needed. We believe this framework may help to bridge the gap from programmers to clinicians and enable them to quickly experiment with their datasets in a reproducible way.
Materials and Methods
This analysis is divided into two main parts: Section Framework describes design principles that we followed while designing AutoRadiomics, and Section Experiment provides information on experiments that were performed to evaluate its performance in publicly available tomography imaging datasets.
Framework
AutoRadiomics is an open-source Python package with an embedded web application with an interactive user interface. The framework can be accessed at https://github.com/pwoznicki/AutoRadiomics, where all the details on its development can be found. The framework is built around the standard steps of a radiomics workflow, including image processing, feature extraction, feature selection, dataset rebalancing, ML model selection, training, optimization, and evaluation. The main components of the framework are presented schematically in Figure 1. AutoRadiomics uses standard libraries validated in multiple radiomics studies, such as pyradiomics (1) for feature extraction, scikit-learn (13) for ML models and data splitting, and imbalanced-learn (14) for over-/undersampling. These reliable building blocks further contribute to creating robust and reproducible workflows. The framework is available under the Apache-2.0 License.
Figure 1. Framework components. AutoRadiomics has a modular architecture, and its components are based on the typical steps in a radiomics analysis. *The first analysis step, automatic segmentation, is not performed inside the framework directly, but a script is generated that can be run separately.
Data Preparation and Radiomic Feature Extraction
Data splitting in AutoRadiomics is performed on provided case IDs. Depending on dataset size and application, user can choose to split the data into k folds for cross-validation (with or without a separate test set) or into training/validation/test sets. Radiomic features, including standard shape, intensity, and texture features, are extracted with pyradiomics, with additional parameters specified in the parameter file. A few built-in options are provided for this purpose, including extraction parameters validated in previous studies (12, 15). Additional optimizations in computing resource allocation make the extraction process more efficient.
Hyperparameter Optimization and Experiment Tracking
Hyperparameter optimization is performed using the Optuna framework (16), which dynamically constructs the search space for hyperparameters and automatically chooses optimal ones. The framework simultaneously optimizes the choice and hyperparameters of the ML classifiers as well as feature selection and oversampling methods, which greatly simplifies the training workflow. The following classifiers are included: logistic regression, support vector machines, random forest, and extreme gradient boosting (XGBoost). Experiments are tracked using an integrated MLFlow tracking dashboard, which allows the user to explore the training artifacts as well as the metrics during and after the training process.
Web Application
Recognizing the problems many non-expert users may face when being forced to use a programming interface, we developed a browser application with an interactive user interface on top of the Python package. The app can be run locally as a Docker container, satisfying the necessary privacy concerns. It adopts a straightforward, modular structure to the radiomics workflow and covers sequentially all steps of the analysis pipeline. The output of each intermediary step, training parameters, and logs are stored in the experiment's directory. That enables the user to later come back to the experiment and document the workflow. Figure 2 presents exemplary screenshots of the app. The app also provides utilities for generating Python code that can be then executed as a separate script to perform automatic segmentation using the state-of-the-art nnU-Net framework (17), and for generating radiomics maps using voxel-based feature extraction.
Figure 2. Exemplary screenshots of the web application. The application enables users to perform all the analysis steps including feature extraction, model training, and evaluation, using standardized or custom settings.
Experiment
Data Sources
To validate the developed framework, we used eight datasets from two different sources. Firstly, we used six public datasets from the recently published WORC database (18), which includes multi-institutional annotated CT and MRI datasets with varying clinical applications. The respective classification tasks were (1) well-differentiated liposarcoma vs. lipoma, (2) desmoid-type fibromatosis vs. extremity soft-tissue sarcoma, (3) primary solid liver tumor, malignant vs. benign, (4) gastrointestinal stromal tumor (GIST) vs. intra-abdominal gastrointestinal tumor, (5) colorectal liver metastases vs. non-metastatic tumor, and (6) lung metastases of melanoma vs. lung tumor of different etiology. The database was released together with benchmark results to facilitate reproducibility in the radiomics field and, to our knowledge, we are the first ones to replicate the previously published results (2).
Additionally, two public prostate MRI datasets, which are available on The Cancer Imaging Archive, were used: Prostate MRI and Ultrasound With Pathology and Coordinates of Tracked Biopsy (19) from the University of California, Los Angeles (UCLA) (further referred to as Prostate-UCLA) and PROSTATEx (20) with annotations from Cuocolo et al. (21). These two datasets were selected since they both had segmentations of prostate gland and lesions as well as biopsy evaluation including Gleason Score (GS) available. All lesions from the Prostate-UCLA dataset had targeted biopsy performed. For PROSTATEx, all lesions with PI-RADS ≥3 were biopsied. We trained radiomics models based on either the whole prostate gland or the target lesion masks in T2-weighted MR images to differentiate between benign prostate lesions and prostate cancer, as well as between clinically significant and clinically insignificant prostate cancer.
Data Processing
The study flowchart is presented in Figure 3. For each dataset, we split 80% of the data into training and 20% into the test set. Then, we split the training set into 5 folds to perform hyperparameter optimization using a cross-validation approach. Image and segmentation data were converted into the NIfTI format, where necessary, and no additional image preprocessing was applied. For feature extraction, we used separate extraction and image processing parameter sets for MRI and CT datasets, as recommended by the IBSI (3). Hyperparameter optimization was performed for each dataset with Optuna using 200 trials of the Tree-structured Parzen Estimator (TPE) algorithm to maximize the objective function.
Statistical Analysis
Receiver operating characteristic (ROC) curves were generated for each independent variable and the area under the curve (AUC) was calculated. The diagnostic efficacy of the model was additionally evaluated using the F1 score, sensitivity, and specificity, and was reported with 95% confidence intervals (95% CI) obtained with the bootstrap technique. The bootstrap used 1,000 resamples (with replacement) of predicted probabilities to determine the 95% CI. All analyses were performed with the AutoRadiomics framework, using Python 3.8.10.
Results
All the experiments were successfully implemented using Python, but can also be reproduced using the interactive web application. Supplementary Figure S1 shows the code extract required to run the optimization and evaluation process for a selected dataset (the implementation assumes a table with data paths is already created). The optimal configurations of models selected for each task are presented in Supplementary Appendix S3. The execution time of the whole pipeline, including the optimization, took around 1 h on a machine with 16 GB RAM and 8-core AMD Ryzen 5800X processor.
The details of training and test cohorts for each task are shown in Table 1. In total, we included 1895 patients in our analyses. In the six datasets from the WORC database, the class distribution was approximately balanced. For the two prostate datasets, the distribution of classes differed between datasets: in PROSTATEx, 50% of index lesions were classified as benign, 15% as GS 6, and 35% as GS ≥7, compared to only 23% of index lesions classified as benign, 24% as GS = 6, and 54% as GS≥7 in Prostate-UCLA.
Table 2 summarizes the classification results and Figure 4 presents the corresponding ROC curves for all included datasets. In the following, we report the results from the test cohorts.
In the WORC database, we obtained results ranging from weak discrimination for the Melanoma dataset (AUC = 0.56 [95% CI: 0.29–0.93], F1 = 0.48 [95% CI: 0.15–0.76]) to excellent discrimination for the Lipo dataset (AUC = 0.93 [95% CI: 0.77–1.0], F1 = 0.85 [95% CI: 0.67–1.0]) and the Desmoid dataset (AUC = 0.90 [95% CI: 0.79–0.98], F1 = 0.77 [95% CI: 0.57–0.92]).
For prostate datasets, results are reported separately for classification using features from either prostate or lesion masks. The discrimination was acceptable for both prostate masks (AUC = 0.73 [95% CI: 0.55–0.89], F1 = 0.73 [0.55–0.86]) as well as lesion masks (AUC = 0.72 [95% CI: 0.56–0.86], F1 = 0.68 [95% CI: 0.48–0.84]) in the PROSTATEx dataset, and moderate for prostate masks (AUC = 0.61 [95% CI: 0.51–0.71], F1 = 0.78 [95% CI: 0.72–0.74]) and lesion masks (AUC = 0.65 [95% CI: 0.55–0.74], 0.73 [95% CI: 0.67–0.80]) in the Prostate-UCLA dataset. Both prostate datasets were additionally validated using the other dataset, and their performance varied from AUC = 0.51 [95% CI: 0.41–0.61] for the PROSTATEx model using lesion masks evaluated in Prostate-UCLA to AUC = 0.77 [95% CI: 0.60–0.91] for the Prostate-UCLA using lesion masks evaluated in PROSTATEx.
The additional evaluation of the prostate MRI datasets for differentiation between clinically significant and clinically non-significant prostate cancer is presented in the Supplementary Table S2. For this challenging task, the results were worse than those for prostate cancer detection, with AUCs ranging from 0.40 [95% CI: 0.29–0.50] for the Prostate-UCLA dataset to AUC = 0.70 [95% CI: 0.33–0.97] for the PROSTATEx dataset, trained with prostate masks. The external validation results in this dataset showed AUCs in the range of 0.37 to 0.70 with high variability.
Discussion
In this study, we introduced and validated a new open-source, interactive framework for reproducible radiomics research. The tool aids in selecting the optimal model for a given task, and the associated web application lowers the entry threshold for clinicians who want to contribute to the field of radiomics research and foster clinical translation.
We evaluated AutoRadiomics in six different classification tasks from the WORC database. It achieved consistently high AUCs in both cross-validation and the test set, in the direct comparison of our results vs. those reported in the original publication on the dataset (2): 0.93 vs. 0.83 for the Lipo dataset, 0.90 vs. 0.82 for the Desmoid dataset, 0.67 vs. 0.81 for the Liver dataset, 0.69 vs. 0.77 for the GIST dataset, 0.75 vs. 0.68 for the CRLM dataset, 0.56 vs. 0.51 for the Melanoma dataset. That means, our framework achieved comparable results, higher in 4/6 tasks, and lower in 2/6 tasks. We also evaluated our framework in two public prostate MRI datasets, achieving AUCs in the range of 0.61–0.73 for internal validation, and 0.51–0.77 for external validation. Those results prove that AutoRadiomics can be successfully applied off-the-shelf and achieve competitive results with its automatic configuration. We believe the differences between ours and previously reported results may be largely explained by the relatively small sample sizes, different data splitting, and differing choice of classifiers. It has to be noted, however, that, similarly to Starmans et al. (2), we achieved best results for the Lipo dataset, and worst for the Melanoma dataset, which suggests both approaches have converged to an optimal solution.
Quantitative evaluation of disease patterns in medical images which are invisible to the human eye has shown diagnostic potential in multiple retrospective studies, but large-scale clinical validation and adoption are still missing (22). We believe that an accessible toolkit for exploratory data analysis and a standardized workflow is a key component in developing the field toward clinical translation. With this in mind, we released AutoRadiomics as an intuitive open-source framework that structures the radiomics workflow and makes it more accessible and reproducible.
Recent advances in automated ML have the potential to empower healthcare professionals with limited data science expertise (23). Inspired by those breakthroughs, new platforms for ML applied to medical imaging have recently been introduced, such as WORC (2), which focuses on automatic construction and optimization of the radiomics workflow. While this platform also provides an automated solution and is very extensive in scope, AutoRadiomics sets itself apart with its interactive web interface, state-of-the-art tooling, and additional utilities (i.e., for visualization and segmentation).
With our web application, we hope to shift the focus from metrics to interpretability, which is achieved through comprehensive visualizations and radiomics maps. We would like to point out a few scenarios, where AutoRadiomics could be especially helpful: (1) for clinicians exploring their dataset using the embedded web application to gain quick insight into their data, (2) for researchers using Python for radiomic analyses, who want to complement their current workflow or add a benchmark or reference standard, (3) for an inter-institutional collaboration as means of facilitating results sharing and workflow reproducibility.
Currently, our framework can be used only for binary classification tasks, which limits its applicability. We are planning to extend it in the future to handle multiclass classification, regression tasks, and survival data. Furthermore, some processing steps such as automatic segmentation using deep learning require GPU capability, which is why it is not integrated into our framework and only the code for performing segmentation with a nnU-Net can be generated. AutoRadiomics does not require a powerful GPU and a modern personal computer is enough to run it. One should also keep in mind that the results of any optimized model have to be considered with caution and no abstraction layer (such as our web application) may replace true expert domain knowledge.
In conclusion, we herein presented AutoRadiomics, a framework for intuitive and reproducible radiomics research. We described its key features as well as the underlying architecture, and we discuss its most promising use cases. Finally, we validated it extensively in eight public datasets to show its consistently high performance in various and diverse classification tasks. We believe that AutoRadiomics may help to improve the quality and reproducibility of future radiomics studies, and, through its accessible interface, may bring those studies closer to clinical translation.
Data Availability Statement
The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author/s.
Ethics Statement
Ethical review and approval was not required for this study in accordance with the local legislation and institutional requirements.
Author Contributions
Study conception: BB. Radiomics analysis: PW. Statistical analysis: PW and FL. Draft writing: PW and BB. Manuscript edition: FL and TB. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) (SPP 2177 to BB).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fradi.2022.919133/full#supplementary-material
References
1. van Griethuysen JJM, Fedorov A, Parmar C, Hosny A, Aucoin N, Narayan V, et al. Computational radiomics system to decode the radiographic phenotype. Cancer Res. (2017) 77:e104–7. doi: 10.1158/0008-5472.CAN-17-0339
2. Starmans MPA, van der Voort SR, Phil T, Timbergen MJM, Vos M, Padmos GA, et al. Reproducible radiomics through automated machine learning validated on twelve clinical applications [Internet]. arXiv [eess.IV]. (2021). Available online at: http://arxiv.org/abs/2108.08618
3. Zwanenburg A, Leger S, Vallières M, Löck S. Image biomarker standardisation initiative. Radiology. (2020) 295:328–38. doi: 10.1148/radiol.2020191145
4. Lambin P, Leijenaar RTH, Deist TM, Peerlings J, Jong EEC de, van Timmeren J, et al. Radiomics: the bridge between medical imaging and personalized medicine. Nat Rev Clin Oncol. (2017) 14:749–62. doi: 10.1038/nrclinonc.2017.141
5. Lambin P, Rios-Velazquez E, Leijenaar R, Carvalho S, van Stiphout RGPM, Granton P, et al. Radiomics: Extracting more information from medical images using advanced feature analysis. Eur J Cancer. (2012) 48:441–6. doi: 10.1016/j.ejca.2011.11.036
6. Baessler B, Luecke C, Lurz J, Klingel K, Roeder M, Waha S, et al. Cardiac MRI Texture Analysis of T1 and T2 maps in patients with infarctlike acute myocarditis. Radiology. (2018) 289:357–65. doi: 10.1148/radiol.2018180411
7. Liu Q, Li J, Liu F, Yang W, Ding J, Chen W, et al. A radiomics nomogram for the prediction of overall survival in patients with hepatocellular carcinoma after hepatectomy. Cancer Imaging. (2020) 20:82. doi: 10.1186/s40644-020-00360-9
8. Aerts HJWL, Velazquez ER, Leijenaar RTH, Parmar C, Grossmann P, Carvalho S, et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat Commun. (2020) 5:1–9. doi: 10.1038/ncomms5644
9. Bera K, Braman N, Gupta A, Velcheti V, Madabhushi A. Predicting cancer outcomes with radiomics and artificial intelligence in radiology. Nat Rev Clin Oncol. (2022) 19:132–46. doi: 10.1038/s41571-021-00560-7
10. Blüthgen C, Patella M, Euler A, Baessler B, Martini K, von Spiczak J, et al. Computed tomography radiomics for the prediction of thymic epithelial tumor histology, TNM stage and myasthenia gravis. PLoS ONE. (2021) 16:e0261401. doi: 10.1371/journal.pone.0261401
11. Mühlbauer J, Kriegmair MC, Schöning L, Egen L, Kowalewski K-F, Westhoff N, et al. Value of radiomics of perinephric fat for prediction of intraoperative complexity in renal tumor surgery. Urol Int. (2021) 1–12. doi: 10.1159/000520445
12. Baessler B, Nestler T, Pinto dos Santos D, Paffenholz P, Zeuch V, Pfister D, et al. Radiomics allows for detection of benign and malignant histopathology in patients with metastatic testicular germ cell tumors prior to post-chemotherapy retroperitoneal lymph node dissection. Eur Radiol. (2020) 30:2334–45. doi: 10.1007/s00330-019-06495-z
13. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine Learning in Python. J Mach Learn Res. (2011) 12:2825–30. doi: 10.5555/1953048.2078195
14. Lemaître G, Nogueira F, Aridas CK. Imbalanced-learn: a python toolbox to tackle the curse of imbalanced datasets in machine learning. J Mach Learn Res. (2017) 18:1–5. doi: 10.5555/3122009.3122026
15. Woznicki P, Westhoff N, Huber T, Riffel P, Froelich MF, Gresser E, et al. Multiparametric MRI for prostate cancer characterization: combined use of radiomics model with PI-RADS and clinical parameters. Cancers. (2020) 12:1767. doi: 10.3390/cancers12071767
16. Akiba T, Sano S, Yanase T, Ohta T, Koyama M. Optuna: A Next-generation Hyperparameter Optimization Framework. arXiv:190710902 [cs, stat]. (2019). Available from: http://arxiv.org/abs/1907.10902 (cited April 4, 2022).
17. Baumgartner M, Jaeger PF, Isensee F, Maier-Hein KH. nnDetection: A Self-configuring Method for Medical Object Detection. arXiv:210600817 [cs.eess]. (2021) 12905:530–9. doi: 10.1007/978-3-030-87240-3_51
18. Starmans MPA, Timbergen MJM, Vos M, Padmos GA, Grünhagen DJ, Verhoef C, et al. The WORC database: MRI and CT scans, segmentations, and clinical labels for 930 patients from six radiomics studies [Internet]. medRxiv. (2021). Available from: https://www.medrxiv.org/content/10.1101/2021.08.19.21262238v1 (cited March 27, 2022).
19. Sonn GA, Natarajan S, Margolis DJA, MacAiran M, Lieu P, Huang J, et al. Targeted biopsy in the detection of prostate cancer using an office based magnetic resonance ultrasound fusion device. J Urol. (2013) 189:86–92. doi: 10.1016/j.juro.2012.08.095
20. Litjens G, Debats O, Barentsz J, Karssemeijer N, Huisman H, TCIA Team. SPIE-AAPM PROSTATEx Challenge Data [Internet]. The Cancer Imaging Archive. (2017). Available from: https://wiki.cancerimagingarchive.net/x/iIFpAQ (cited March 31, 2022).
21. Cuocolo R, Stanzione A, Castaldo A, Lucia DRD, Imbriaco M. Quality control and whole-gland, zonal and lesion annotations for the PROSTATEx challenge public dataset. Eur J Radiol. (2021). p. 138. Available from: https://www.ejradiology.com/article/S0720-048X(21)00127-3/fulltext (cited April 4, 2022).
22. Pinto dos Santos D, Dietzel M, Baessler B. A decade of radiomics research: are images really data or just patterns in the noise? Eur Radiol. (2021) 31:1–4. doi: 10.1007/s00330-020-07108-w
Keywords: radiomics, radiology, machine learning, reproducibility, workflow, image analysis
Citation: Woznicki P, Laqua F, Bley T and Baeßler B (2022) AutoRadiomics: A Framework for Reproducible Radiomics Research. Front. Radiol. 2:919133. doi: 10.3389/fradi.2022.919133
Received: 13 April 2022; Accepted: 20 June 2022;
Published: 07 July 2022.
Edited by:
Alexandra Wennberg, Karolinska Institutet (KI), SwedenReviewed by:
Loïc Duron, Fondation Adolphe de Rothschild, FranceShahriar Faghani, Mayo Clinic, United States
Copyright © 2022 Woznicki, Laqua, Bley and Baeßler. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Piotr Woznicki, woznicki_p@ukw.de