Araştırma Makalesi

A Supervised Learning Approach With Residual Attention Connections

Cilt: 5 Sayı: 1 21 Haziran 2024
Ali Hamza , Fazal Muhammad , Talha Ali , Fazal E-wahab , Muhammad Ismail *
PDF İndir
EN TR

A Supervised Learning Approach With Residual Attention Connections

Abstract

Our study aims to improve speech quality despite background noise, which often disrupts clear communication. We focus on developing efficient and effective models that work well on devices with limited resources. We draw inspiration from computational auditory scene analysis techniques to train our models to differentiate speech from background noise while keeping computational demands low. We introduce two models: CRN-WRC (Convolutional Recurrent Network without Residual Connections) and CRN-RCAG (Convolutional Recurrent Network with Residual Connections and Attention Gates). Our thorough testing shows that our models significantly enhance speech quality and understanding, even in noisy environments with varying background noise levels. Notably, the CRN-RCAG model consistently outperforms the CRN-WRC, particularly in handling untrained noise types. We achieve impressive results by integrating residual connections and attention gates into our models while maintaining computational efficiency.

Keywords

 Speech enhancement ,  Convolutional Recurrent Network ,  supervised learning

Kaynakça

  1. [1] Kheddar, Hamza, et al. "Deep transfer learning for automatic speech recognition: Towards better generalization." Knowledge-Based Systems 277 (2023): 110851.
  2. [2] Kwak, Chanbeom, and Woojae Han. "Towards size of scene in auditory scene analysis: A systematic review." Journal of Audiology & Otology 24.1 (2020): 1.
  3. [3] Wang, DeLiang, and Jitong Chen. "Supervised speech separation based on deep learning: An overview." IEEE/ACM transactions on audio, speech, and language processing 26.10 (2018): 1702-1726.
  4. [4] Nossier, Soha A., et al. "Mapping and masking targets comparison using different deep learning based speech enhancement architectures." 2020 International Joint Conference on Neural Networks (IJCNN). IEEE, 2020.
  5. [5] Ye, Zhongfu, Nasir Saleem, and Hamza Ali. "Efficient gated convolutional recurrent neural networks for real-time speech enhancement." (2023). [6] Hsieh, Tsun-An, et al. "Wavecrn: An efficient convolutional recurrent neural network for end-to-end speech enhancement." IEEE Signal Processing Letters 27 (2020): 2149-2153.
  6. [7] Wang, Kai. Novel Deep Learning Approaches for Single-Channel Speech Enhancement. Diss. Concordia University, 2022.
  7. [8] Haar, Lynn Vonder, Timothy Elvira, and Omar Ochoa. "An analysis of explainability methods for convolutional neural networks." Engineering Applications of Artificial Intelligence 117 (2023): 105606.
  8. [9] Le, Xiaohuai, et al. "DPCRN: Dual-path convolution recurrent network for single channel speech enhancement." arXiv preprint arXiv:2107.05429 (2021).
  9. [10] Marcu, David C., and Cristian Grava. "The impact of activation functions on training and performance of a deep neural network." 2021 16th International Conference on Engineering of Modern Electric Systems (EMES). IEEE, 2021.
  10. [11] Ye, Zhongfu, Nasir Saleem, and Hamza Ali. "Efficient gated convolutional recurrent neural networks for real-time speech enhancement." (2023).

Kaynak Göster

APA
Hamza, A., Muhammad, F., Ali, T., E-wahab, F., & Ismail, M. (2024). A Supervised Learning Approach With Residual Attention Connections. Journal of Science, Technology and Engineering Research, 5(1), 78-85. https://doi.org/10.53525/jster.1469477
AMA
1.Hamza A, Muhammad F, Ali T, E-wahab F, Ismail M. A Supervised Learning Approach With Residual Attention Connections. Journal of Science, Technology and Engineering Research. 2024;5(1):78-85. doi:10.53525/jster.1469477
Chicago
Hamza, Ali, Fazal Muhammad, Talha Ali, Fazal E-wahab, ve Muhammad Ismail. 2024. “A Supervised Learning Approach With Residual Attention Connections”. Journal of Science, Technology and Engineering Research 5 (1): 78-85. https://doi.org/10.53525/jster.1469477.
EndNote
Hamza A, Muhammad F, Ali T, E-wahab F, Ismail M (01 Haziran 2024) A Supervised Learning Approach With Residual Attention Connections. Journal of Science, Technology and Engineering Research 5 1 78–85.
IEEE
[1]A. Hamza, F. Muhammad, T. Ali, F. E-wahab, ve M. Ismail, “A Supervised Learning Approach With Residual Attention Connections”, Journal of Science, Technology and Engineering Research, c. 5, sy 1, ss. 78–85, Haz. 2024, doi: 10.53525/jster.1469477.
ISNAD
Hamza, Ali - Muhammad, Fazal - Ali, Talha - E-wahab, Fazal - Ismail, Muhammad. “A Supervised Learning Approach With Residual Attention Connections”. Journal of Science, Technology and Engineering Research 5/1 (01 Haziran 2024): 78-85. https://doi.org/10.53525/jster.1469477.
JAMA
1.Hamza A, Muhammad F, Ali T, E-wahab F, Ismail M. A Supervised Learning Approach With Residual Attention Connections. Journal of Science, Technology and Engineering Research. 2024;5:78–85.
MLA
Hamza, Ali, vd. “A Supervised Learning Approach With Residual Attention Connections”. Journal of Science, Technology and Engineering Research, c. 5, sy 1, Haziran 2024, ss. 78-85, doi:10.53525/jster.1469477.
Vancouver
1.Ali Hamza, Fazal Muhammad, Talha Ali, Fazal E-wahab, Muhammad Ismail. A Supervised Learning Approach With Residual Attention Connections. Journal of Science, Technology and Engineering Research. 01 Haziran 2024;5(1):78-85. doi:10.53525/jster.1469477