Speech Noise Reduction via Intelligent Spectral Gain Selection and Modification

Authors

  • Huiting Yu, Anton Louise De Ocampo, Rowell Hernandez

Keywords:

attention mechanism, deep learning, noise suppression, speech processing

Abstract

Speech information is distributed across the frequency band of 300 Hz to about 3000-4000 Hz. Noise affects speech information distinctively between different spectral components of the speech signal. This means that noise infusion to individual spectral components is variable when measured in signal-to-noise ratio. This variation can be picked up and efficiently discerned by the workings of the human brain. The human brain can perform spectral selection and gain modification to address the various effects of noise in a speech signal. When faced with noise, the brain combines top-down and bottom-up processing to improve the signal-to-noise ratio of the speech signal. The brain's ability to select spectra with more information is not utilized in existing or traditional methods. This paper proposes a novel deep neural network-based method for voice noise reduction. The formulations are determined by attention-based neural networks using spectral gain adjustments as the basis for the proposed technique. The intelligent spectral gain selection and modification using attention mechanisms are introduced after speech signal preprocessing. The noisy signal passes through spectral decomposition where each component has assigned weights based on the proposed attention network. The work shows how the addition of spectral gain adjustments affects the suppression of noise in voice signals by an attention-network-based algorithm. The attention network can focus on the spectral components that are most informative by suppressing bands where noise is excessive. Based on the identified weights, the spectral component is attenuated or amplified to extract the most information content possible. The proposed framework obtained an SNR and segmented SNR of 7.1138 and 1.9950 respectively, higher than existing methods.

Downloads

Download data is not yet available.

References

X. Wang and L. Xu, “Speech perception in noise: Masking and unmasking,” Journal of Otology, vol. 16, no. 2, pp. 109–119, 2021.

A. de Cheveigné, “Harmonic cancellation—A fundamental of auditory scene analysis,” Trends in Hearing, vol. 25, p. 23312165211041424, 2021.

Mahadevaswamy and D. Ravi, “Robust perceptual wavelet packet features for recognition of continuous Kannada speech,” Wireless Personal Communications, vol. 121, no. 3, pp. 1781–1804, 2021.

M. S. Islam, Y. Zhu, M. I. Hossain, R. Ullah, and Z. Ye, “Supervised single channel dual domains speech enhancement using sparse non-negative matrix factorization,” Digital Signal Processing, vol. 100, p. 102697, 2020.

M. Iqbal, S. A. Raza, M. Abid, F. Majeed, and A. A. Hussain, “Artificial neural network based emotion classification and recognition from speech,” International Journal of Advanced Computer Science and Applications, vol. 11, no. 12, 2020.

M. A. A. Albadr, S. Tiun, M. Ayob, F. T. AL-Dhief, K. Omar, and M. K. Maen, “Speech emotion recognition using optimized genetic algorithm-extreme learning machine,” Multimedia Tools and Applications, vol. 81, no. 17, pp. 23963–23989, 2022.

M. Une and R. Miyazaki, “Evaluation of sound quality and speech recognition performance using harmonic regeneration for various noise reduction techniques,” in RISP Int. Workshop Nonlinear Circuits, Commun Signal Process.(NCSP), 2017, pp. 377–380.

H. H. Nuha and A. A. Absa, “Noise reduction and speech enhancement using wiener filter,” in 2022 International Conference on Data Science and Its Applications (ICoDSA), 2022, pp. 177–180.

C. I. Muresan, I. R. Birs, E. H. Dulf, D. Copot, and L. Miclea, “A review of recent advances in fractional-order sensing and filtering techniques,” Sensors, vol. 21, no. 17, p. 5920, 2021.

J. S. Jakati and S. S. Kuntoji, “A noise reduction method based on modified LMS algorithm of real time speech signals,” WSEAS Trans Environ Dev, vol. 16, no. 13, 2021.

T. P. Zieliński and T. P. Zieliński, “FIR Adaptive Filters,” Starting Digital Signal Processing in Telecommunication Engineering: A Laboratory-based Course, pp. 317–343, 2021.

X. H. Xie, L. N. Zhou, and Y. J. Xie, “Design and Simulation of Active Noise Cancelling Earphone System Based on FXLMS Algorithm,” in 2022 4th International Conference on Natural Language Processing (ICNLP), 2022, pp. 626–630.

O. Barkovska, V. Kholiev, and V. Lytvynenko, “Study of noise reduction methods in the sound sequence when solving the speech-to-text problem,” Advanced Information Systems, vol. 6, no. 1, pp. 48–54, 2022.

S. C. Venkateswarlu, N. U. Kumar, and A. Karthik, “Speech enhancement using recursive least square based on real-time adaptive filtering algorithm,” in 2021 6th International Conference for Convergence in Technology (I2CT), 2021, pp. 1–4.

A. Chiheb and H. Khelladi, “Performance Comparison of LMS and RLS Algorithms for Ambient Noise Attenuation,” International Journal of Electrical and Computer Engineering Research, vol. 4, no. 1, pp. 14–19, 2024.

S. Olika and A. Rajani, “Adaptive noise cancellation for speech signal,” Int J Sci Res Publ, vol. 10, no. 9, 2020.

J. Blessy and C. S. Christopher, “A Survey on Filtering Methods Used to Remove Noise in Speech and Music Signal,” in 2022 6th International Conference on Electronics, Communication and Aerospace Technology, 2022, pp. 140–145.

Downloads

Published

14.08.2024

How to Cite

Huiting Yu. (2024). Speech Noise Reduction via Intelligent Spectral Gain Selection and Modification. International Journal of Intelligent Systems and Applications in Engineering, 12(4), 2642 –. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/6691

Issue

Section

Research Article