| dc.rights.license | Visos teisės saugomos / All rights reserved | en_US |
| dc.contributor.author | Yushchenko, Artur | |
| dc.contributor.author | Smelyakov, Kirill | |
| dc.contributor.author | Chupryna, Anastasiya | |
| dc.date.accessioned | 2026-01-09T11:32:31Z | |
| dc.date.available | 2026-01-09T11:32:31Z | |
| dc.date.issued | 2025 | |
| dc.identifier.isbn | 9798331598747 | en_US |
| dc.identifier.issn | 2831-5634 | en_US |
| dc.identifier.uri | https://etalpykla.vilniustech.lt/handle/123456789/159709 | |
| dc.description.abstract | This paper explores three prominent deep learning architectures — Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), and Vision Transformers (ViT) — for emotion recognition, examining their potential strengths and weaknesses under various conditions. It discusses how each approach may capture critical spatial, temporal, or global features in emotional data, highlighting differences in feature extraction, representational capacity, and scalability. Additionally, new solutions are proposed to enhance accuracy and adaptability, integrating design principles that address recognized challenges in real-world implementations. Novel insights are offered on aligning model selection with specific application demands, such as the nature of input signals, available computational resources, and desired real-time performance. While the comparative analysis remains broad to accommodate diverse use cases, it underscores the importance of carefully balancing accuracy and efficiency. Conclusions drawn from the investigation include recommendations on when each architecture may be most advantageous, providing a flexible framework for researchers and practitioners to navigate the trade-offs. These findings have implications for developing adaptive emotion recognition systems that leverage state-of-the-art deep learning techniques across multiple contexts. | en_US |
| dc.format.extent | 6 p. | en_US |
| dc.format.medium | Tekstas / Text | en_US |
| dc.language.iso | en | en_US |
| dc.relation.uri | https://etalpykla.vilniustech.lt/handle/123456789/159405 | en_US |
| dc.source.uri | https://ieeexplore.ieee.org/document/11016864 | en_US |
| dc.subject | emotion recognition | en_US |
| dc.subject | deep learning | en_US |
| dc.subject | convolutional neural networks | en_US |
| dc.subject | recurrent neural networks | en_US |
| dc.subject | vision trans-formers | en_US |
| dc.subject | facial expression analysis | en_US |
| dc.title | Evaluating CNN, RNN, and Vision Transformer for Emotion Recognition: Strengths and Weaknesses | en_US |
| dc.type | Konferencijos publikacija / Conference paper | |
| dcterms.accrualMethod | Rankinis pateikimas / Manual submission | en_US |
| dcterms.issued | 2025-06-02 | |
| dcterms.references | 21 | en_US |
| dc.description.version | Taip / Yes | en_US |
| dc.contributor.institution | Kharkiv National University of Radio Electronics | en_US |
| dcterms.sourcetitle | 2025 IEEE Open Conference of Electrical, Electronic and Information Sciences (eStream), April 24, 2025, Vilnius, Lithuania | en_US |
| dc.identifier.eisbn | 9798331598730 | en_US |
| dc.identifier.eissn | 2690-8506 | en_US |
| dc.publisher.name | IEEE | en_US |
| dc.publisher.country | United States of America | en_US |
| dc.publisher.city | New York | en_US |
| dc.identifier.doi | https://doi.org/10.1109/eStream66938.2025.11016864 | en_US |