TR
EN
Application of Emotion Analysis in Deep Learning Techniques
Abstract
Emotion recognition has become a pivotal technology in advancing human-computer interaction with applications spanning fields such as healthcare, entertainment, and customer experience. This paper evaluates the performance of five deep learning models—YOLOv8m-cls, ResNet50, EfficientNetB5, MobileNetV2, and DenseNet121—in detecting emotions from facial expressions. Leveraging the AffectNet dataset, which initially contained eight emotional categories, we focused on five emotions after excluding three due to low data availability and similarity. The emotions processed include anger, happiness, sadness, surprise, and fear. The models were fine-tuned through transfer learning, demonstrating that YOLOv8m-cls performed best, balancing accuracy, speed, and generalization, making it suitable for real-time applications. ResNet50 and EfficientNetB5 also performed well, with ResNet50 excelling in handling complex facial features and EfficientNetB5 offering computational efficiency with high accuracy. The study also highlights challenges such as intra-class variability and inter-class similarity, which continue to affect model performance. These findings underscore the importance of selecting model architectures based on specific application requirements and suggest that future research should explore integrating multimodal data to enhance emotion recognition systems.
Keywords
Thanks
Makalemizi inceleyecek olan değerli Editör ve Hakem hocalarımıza şimdiden teşekkür ederiz.
References
- [1] P. Ekman, “Facial expression and emotion,” Am. Psychol, vol. 48, no. 4, pp. 384–392, 1993. DOI:10.1037/0003-066X.48.4.384
- [2] M. Pantic and L. J. Rothkrantz, “An expert system for recognition of facial actions and their intensity,” AAAI/IAAI, pp. 1026–1033, 2000.
- [3] Z. Zeng, M. Pantic, G. I. Roisman and T. S. Huang, “A survey of affect recognition methods: Audio, visual and spontaneous expressions,” Proceedings of the 9th International Conference on Multimodal Interfaces, pp. 126–133, 2007.
- [4] J. Zhang, Z. Yin, P. Cheng and S. Nichele, “Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review,” Inf Fusion, 2020. DOI:10.1016/j.inffus.2020.01.011
- [5] Y. Tian, T. Kanade and J. F. Cohn, “Facial expression recognition,” in Handbook of Face Recognition, S. Z. Li and A. K. Jain, Eds., Springer, 2011, pp. 487–519. DOI:10.1007/978-0-85729-932-1_19
- [6] P. Ekman and W. V. Friesen, Facial Action Coding System. Consulting Psychologists Press, 1978.
- [7] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar and I. Matthews, “The extended cohn-kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression,” 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition – Workshops (CVPRW), pp. 94–101, 2010. DOI:10.1109/CVPRW.2010. 5543262
- [8] G. Tolias and O. Chum, “Asymmetric feature maps with application to sketch-based retrieval,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3155–3164, 2017.
Details
Primary Language
English
Subjects
Image Processing , Deep Learning
Journal Section
Research Article
Publication Date
December 30, 2025
Submission Date
August 3, 2025
Acceptance Date
November 25, 2025
Published in Issue
Year 2025 Volume: 16 Number: 4