DECODING BRAIN SIGNALS FOR INNER SPEECH: EEG SIGNAL SUB-BAND DECOMPOSITION AND MACHINE LEARNING APPROACH
Abstract
Decoding inner speech is a rapidly advancing field with significant implications for brain-computer interfaces (BCIs) and assistive technologies. In this study how sub-band decomposition, feature types, and classifier choice affect inner-speech classification were assessed. A comprehensive comparative framework is presented, employing various sub-band decomposition techniques including Tunable Q-Factor Wavelet Transform, Empirical Mode Decomposition, Variational Mode Decomposition, and Circulant Singular Spectrum Analysis. From the decomposed signals, 17 statistical, 4 entropy, and 5 frequency features were extracted individually and in hybrid combinations before reduction with Least Absolute Shrinkage and Selection Operator. Classification performance was evaluated with 10-fold cross-validation, using Support Vector Machine, Feedforward Neural Network, and K-Nearest Neighbor. Performance was summarized as the mean across 30 independent runs per condition over six metrics. A three-way repeated measures ANOVA with Greenhouse-Geisser correction showed significant main effects and interactions across all metrics (p < 0.001), with large effect sizes (η2p >= 0.14). The best overall configuration was EMD-Hybrid-SVM outperforming baseline models, yielding a group-level mean accuracy of 54.13%. These findings demonstrate the potential of sub-band decomposition and hybrid features for improving inner speech decoding and contribute to the development of more robust EEG-based BCIs.
Keywords
Brain Computer Interface , Brain Decoding , EEG Classification , Inner Speech , Sub-Band Decomposition
References
- Abdulghani, M. M., Walters, W. L., & Abed, K. H. (2024). Inner Speech Recognition for Mutism and Speech Disorder Using Brain-Computer Interface. Fortune Journal of Health Sciences, 985–995. https://doi.org/10.26502/fjhs.361
- Awuah, W. A., Ahluwalia, A., Darko, K., Sanker, V., Tan, J. K., Tenkorang, P. O., Ben-Jaafar, A., Ranganathan, S., Aderinto, N., & Mehta, A. (2024). Bridging Minds and Machines: The Recent Advances of Brain-Computer Interfaces in Neurological and Neurosurgical Applications. World Neurosurgery, 189, 138–153. https://doi.org/10.1016/j.wneu.2024.05.104
- Borra, D., Fantozzi, S., & Magosso, E. (2019). EEG Motor Execution Decoding via Interpretable Sinc-Convolutional Neural Networks. Mediterranean Conference on Medical and Biological Engineering and Computing, 1113–1122.
- Bueno-López, M., Giraldo, E., & Molinas, M. (2017). Analysis of Neural Activity from EEG Data Based on EMD Frequency Bands. 2017 24th IEEE International Conference on Electronics, Circuits and Systems (ICECS), 401–405.
- Chowdhury, A. T., Pedersen, S., Chowdhury, M. E. H., Al Shibli, A. N., Khanafer, Y., AbuHaweeleh, M. N., & Hassanein, A. (2025). Neural Signals, Machine Learning, and the Future of Inner Speech Recognition. Frontiers in Human Neuroscience, 19, 1637174.
- Cohen, J. (2013). Statistical Power Analysis for the Behavioral Sciences. routledge.
- Duraivel, S., Rahimpour, S., Chiang, C.-H., Trumpis, M., Wang, C., Barth, K., Harward, S. C., Lad, S. P., Friedman, A. H., & Southwell, D. G. (2023). High-Resolution Neural Recordings Improve the Accuracy of Speech Decoding. Nature Communications, 14(1), 6938.
- Gasparini, F., Cazzaniga, E., & Saibene, A. (2022). Inner Speech Recognition through Electroencephalographic Signals. ArXiv Preprint ArXiv:2210.06472.
- Güneş, F. (2024). Inner Voice in Silent Reading and Speaking. The Journal of Limitless Education and Research, 9(1), 1–26. https://doi.org/10.29250/sead.1436684
- Hu, H., Pu, Z., Li, H., Liu, Z., & Wang, P. (2022). Learning Optimal Time-Frequency-Spatial Features by the Cissa-Csp Method for Motor Imagery Eeg Classification. Sensors, 22(21), 8526.