Interpretability and accuracy of machine learning algorithms for biomedical time series analysis – a scoping review

Gespeichert in:
Bibliographische Detailangaben
Titel: Interpretability and accuracy of machine learning algorithms for biomedical time series analysis – a scoping review
Autoren: Alan Jovic, Nikolina Frid, Karla Brkic, Mario Cifrek
Quelle: Biomedical Signal Processing and Control
Volume 110
Verlagsinformationen: Elsevier BV, 2025.
Publikationsjahr: 2025
Schlagwörter: Artificial intelligence, Interpretable machine learning, TECHNICAL SCIENCES. Computing, TEHNIČKE ZNANOSTI. Računarstvo, Biomedical signal processing, Time series analysis, Deep learning, Biomedical time series
Beschreibung: Current research in biomedical time series (BTS) (e.g., ECG, EEG) analysis focuses on applications of various deep learning approaches to improve classification, prediction, or assessment of states and disorders. When trained on sufficiently large datasets, such approaches mostly lead to highly accurate, yet uninterpretable models, sometimes with a possibility for post-hoc explainability. Since high-stake areas such as healthcare warrant model explanations and, where possible, high interpretability in addition to model efficiency, there is nowadays a surprising scarcity of interpretable machine learning models proposed in this field. Although the machine learning community is aware of the need for interpretable machine learning in BTS analysis, the proposed models do not reflect this need. In this scoping review, we considered over 30,000 studies from the Web of Science database, screened nearly 500 studies, and selected over 50 high-quality studies for detailed analysis. These studies focus on interpretable methods, accurate methods, and approaches bridging the two. Most studies analyzed ECG and EEG signals and concentrated on a limited range of applications, including emotion recognition, heart diseases, epilepsy, and motor imagery, reflecting the scarcity of quality public datasets. K-nearest neighbors and decision trees were the most used interpretable methods, while convolutional neural networks with recurrent or attention layers, achieved the highest accuracy. The methods that balance interpretability and accuracy in BTS analysis include advanced generalized additive models and optimization-based approaches for decision trees, rule learning, and linear models. These approaches warrant further studies, as only a few of them were applied in BTS analysis.
Publikationsart: Article
Dateibeschreibung: application/pdf
Sprache: English
ISSN: 1746-8094
DOI: 10.1016/j.bspc.2025.108153
Zugangs-URL: https://repozitorij.fer.unizg.hr/islandora/object/fer:13416/datastream/FILE0
https://urn.nsk.hr/urn:nbn:hr:168:867518
https://repozitorij.fer.unizg.hr/islandora/object/fer:13416
Rights: CC BY NC ND
Dokumentencode: edsair.doi.dedup.....06f5269293323c8d5a3c0069ca6e40ef
Datenbank: OpenAIRE
Beschreibung
Abstract:Current research in biomedical time series (BTS) (e.g., ECG, EEG) analysis focuses on applications of various deep learning approaches to improve classification, prediction, or assessment of states and disorders. When trained on sufficiently large datasets, such approaches mostly lead to highly accurate, yet uninterpretable models, sometimes with a possibility for post-hoc explainability. Since high-stake areas such as healthcare warrant model explanations and, where possible, high interpretability in addition to model efficiency, there is nowadays a surprising scarcity of interpretable machine learning models proposed in this field. Although the machine learning community is aware of the need for interpretable machine learning in BTS analysis, the proposed models do not reflect this need. In this scoping review, we considered over 30,000 studies from the Web of Science database, screened nearly 500 studies, and selected over 50 high-quality studies for detailed analysis. These studies focus on interpretable methods, accurate methods, and approaches bridging the two. Most studies analyzed ECG and EEG signals and concentrated on a limited range of applications, including emotion recognition, heart diseases, epilepsy, and motor imagery, reflecting the scarcity of quality public datasets. K-nearest neighbors and decision trees were the most used interpretable methods, while convolutional neural networks with recurrent or attention layers, achieved the highest accuracy. The methods that balance interpretability and accuracy in BTS analysis include advanced generalized additive models and optimization-based approaches for decision trees, rule learning, and linear models. These approaches warrant further studies, as only a few of them were applied in BTS analysis.
ISSN:17468094
DOI:10.1016/j.bspc.2025.108153