Please use this identifier to cite or link to this item:
http://dx.doi.org/10.25673/122078Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Abbas Ali, Samah | - |
| dc.contributor.other | Mustafa Abbas, Jamal | - |
| dc.date.accessioned | 2026-02-09T10:53:06Z | - |
| dc.date.available | 2026-02-09T10:53:06Z | - |
| dc.date.issued | 2025-08 | - |
| dc.identifier.uri | https://opendata.uni-halle.de//handle/1981185920/124026 | - |
| dc.identifier.uri | http://dx.doi.org/10.25673/122078 | - |
| dc.description.abstract | This study introduces a novel and effective method for Speech Emotion Recognition (SER) that combines Spiking Neural Networks (SNNs), Temporal Attention, and Transformer encoders within a powerful hybrid model. SER is essential for improving human-computer interaction by enabling intelligent systems to effectively recognize emotions from speech. Unlike traditional methods that typically rely on shallow classifiers and manually engineered features, our deep learning-based approach takes full advantage of the energy efficiency of SNNs, the selective focus provided by temporal attention, and the long-range temporal modeling capabilities of Transformer architectures. We thoroughly evaluated the performance of this model on a comprehensive multi-dataset corpus, which included TESS, SAVEE, RAVDESS, and CREMA-D. The model achieved an impressive and consistent accuracy of 98% across all emotion classes. These strong results not only demonstrate the model’s superior effectiveness but also highlight its potential for use in real-time, resource-limited environments. Furthermore, this hybrid approach clearly surpasses existing state-of-the-art SER techniques and offers a reliable foundation for application in real-world affective computing scenarios. | - |
| dc.format.extent | 1 Online-Ressource (7 Seiten) | - |
| dc.language.iso | eng | - |
| dc.rights.uri | https://creativecommons.org/licenses/by-sa/4.0/ | - |
| dc.subject.ddc | DDC::6** Technik, Medizin, angewandte Wissenschaften | - |
| dc.title | A Hybrid Spiking-Attention Transformer Model for Robust and Efficient Speech Emotion Recognition on Multi-Dataset Benchmarks | - |
| local.versionType | publishedVersion | - |
| local.publisher.universityOrInstitution | Hochschule Anhalt | - |
| local.openaccess | true | - |
| dc.identifier.ppn | 1951201000 | - |
| cbs.publication.displayform | 2025 | - |
| local.bibliographicCitation.year | 2025 | - |
| cbs.sru.importDate | 2026-02-09T10:51:57Z | - |
| local.bibliographicCitation | Enthalten in Proceedings of the 13th International Conference on Applied Innovations in IT - Koethen, Germany : Edition Hochschule Anhalt, 2025 | - |
| local.accessrights.dnb | free | - |
| Appears in Collections: | International Conference on Applied Innovations in IT (ICAIIT) | |
Files in This Item:
| File | Size | Format | |
|---|---|---|---|
| 2-8-ICAIIT_2025_13(4).pdf | 1.03 MB | Adobe PDF | View/Open |