Research Article

Application of Emotion Analysis in Deep Learning Techniques

Volume: 16 Number: 4 December 30, 2025
TR EN

Application of Emotion Analysis in Deep Learning Techniques

Abstract

Emotion recognition has become a pivotal technology in advancing human-computer interaction with applications spanning fields such as healthcare, entertainment, and customer experience. This paper evaluates the performance of five deep learning models—YOLOv8m-cls, ResNet50, EfficientNetB5, MobileNetV2, and DenseNet121—in detecting emotions from facial expressions. Leveraging the AffectNet dataset, which initially contained eight emotional categories, we focused on five emotions after excluding three due to low data availability and similarity. The emotions processed include anger, happiness, sadness, surprise, and fear. The models were fine-tuned through transfer learning, demonstrating that YOLOv8m-cls performed best, balancing accuracy, speed, and generalization, making it suitable for real-time applications. ResNet50 and EfficientNetB5 also performed well, with ResNet50 excelling in handling complex facial features and EfficientNetB5 offering computational efficiency with high accuracy. The study also highlights challenges such as intra-class variability and inter-class similarity, which continue to affect model performance. These findings underscore the importance of selecting model architectures based on specific application requirements and suggest that future research should explore integrating multimodal data to enhance emotion recognition systems.

Keywords

Thanks

Makalemizi inceleyecek olan değerli Editör ve Hakem hocalarımıza şimdiden teşekkür ederiz.

References

  1. [1] P. Ekman, “Facial expression and emotion,” Am. Psychol, vol. 48, no. 4, pp. 384–392, 1993. DOI:10.1037/0003-066X.48.4.384
  2. [2] M. Pantic and L. J. Rothkrantz, “An expert system for recognition of facial actions and their intensity,” AAAI/IAAI, pp. 1026–1033, 2000.
  3. [3] Z. Zeng, M. Pantic, G. I. Roisman and T. S. Huang, “A survey of affect recognition methods: Audio, visual and spontaneous expressions,” Proceedings of the 9th International Conference on Multimodal Interfaces, pp. 126–133, 2007.
  4. [4] J. Zhang, Z. Yin, P. Cheng and S. Nichele, “Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review,” Inf Fusion, 2020. DOI:10.1016/j.inffus.2020.01.011
  5. [5] Y. Tian, T. Kanade and J. F. Cohn, “Facial expression recognition,” in Handbook of Face Recognition, S. Z. Li and A. K. Jain, Eds., Springer, 2011, pp. 487–519. DOI:10.1007/978-0-85729-932-1_19
  6. [6] P. Ekman and W. V. Friesen, Facial Action Coding System. Consulting Psychologists Press, 1978.
  7. [7] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar and I. Matthews, “The extended cohn-kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression,” 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition – Workshops (CVPRW), pp. 94–101, 2010. DOI:10.1109/CVPRW.2010. 5543262
  8. [8] G. Tolias and O. Chum, “Asymmetric feature maps with application to sketch-based retrieval,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3155–3164, 2017.

Details

Primary Language

English

Subjects

Image Processing , Deep Learning

Journal Section

Research Article

Publication Date

December 30, 2025

Submission Date

August 3, 2025

Acceptance Date

November 25, 2025

Published in Issue

Year 2025 Volume: 16 Number: 4

IEEE
[1]M. Uysal, M. F. Demiral, and A. H. Işik, “Application of Emotion Analysis in Deep Learning Techniques”, DUJE, vol. 16, no. 4, pp. 919–935, Dec. 2025, doi: 10.24012/dumf.1757225.