Designing Highly Secured Speaker Identification with Audio Fingerprinting using MODWT and RBFNN
Keywords:
RBFNN-PNCC, HMM-LSTM, MODWT, Speaker identification, Audio fingerprintingAbstract
The process of matching Speech data with database entries is known as Speaker identification. This research processes a Novel technique to speaker recognition by merging neural networks with audio fingerprinting. In comparison to classical models the radial basis function neural network(RBFNN) combined with the hidden markov model(HMM) provides superior results. To facilitate the denoising process for both spoken and unvoiced speech, including ambient noise, the speech data is divided into low and high frequency segements.The maximal overlap discrete wavelet transform (MODWT) is used to denoise the split signals. When it comes to boundaries MODWT is more resilient. In the presence of white noise the Power normalized cepstral coefficients (PNCC) will provide more accurate results than Mel cepstrums.The results are accurately observed using this suggested strategy in relation to SNR.
Downloads
References
Shiqing zhang,Xiaoming Zhao,Bicheng lei,”Spoken emotion recognition using radial basis function Neural Network”,”Advances in computer science,environment,ecoinformatics and education”.pages 437-442.
Ali bounassif,Noha alnazzawi,Ismail shahin,Said A.Salloum,Noor Hindawi,Mohammed Lataifeh,Ashraf Elnagar,”A novel RBFNN-CNN Model for speaker identification in stressful talking environments”,”Human-Computer interactions,May-2022.
R.L.K.Venkateswarlu,R.Vasantha Kumari,G.Vani jayasri,”Speech recognition using radial basis function neural network”,3rd international conference IEEE,2011.
Kumud Arora,Dr.V.P.Vishwa Karma,Dr.Poonam Garg,”Radial basis function neural network trained withvariant spread learning”,”International journal of Engineering research and Technology”,Volume3.Issue9.(sep-2014).
D.Prabhakaran,S.Sriuppili,”Speech processing:MFCC based feature extraction techniques-an investigation”,”Journal of physics”,2021.
B.Vimal,Muthyam surya,darshan,V S sridhar,Asha ashok,”MFCC based audio classification using machine learning”,”12th international conference IEEE,2021.
Shalbbya ali,Dr.Safdar Tanweer,Syed Sibtain Khalid,Dr.Naseem Rao,”Mel Frequency cepstral coefficient :A Review”, “Proceedings of the 2nd International Conference”,ICIDSSD,2020.
Xuechen Liu,Md Sahidullah,Tom kinnunen,”Optimized power normalized cepstral coefficients towards robust deep speaker verification”,”Automatic Speech recognition IEEE,2021.
Ishwar chandra yadav,Gavadhar pradhan,”Pitch and noise normalized acoustic feature for children’s ASR”,”Digital signal processing”,vol-10,feb-2021.
Chanwookim,Member IEEE and Richard M.Stern,”Power Normalized Cepstral Coefficients for robust speech recognition”,”IEEE/ACM transactions on audio speech and language processing”,Vol 24,No7,July-2016.
Selma ozaydin,Iman khalil alank,”Speech enhancement using maximal overlap discrete wavelet transform”,”Gazi university journal of science”,Dec-2018.
Sasikumar gurumoorthy,Naresh babu muppalaneni,G.sandhya Kumari,”EEG signal denoising using Haar transform and maximal overlap discrete wavelet transform(MODWT) for the finding of Epilepsy”,”Etilogies instrumental diagnosis and treatment”Sep-2020.
Davi v.q.rodrigues,Delong zuo,Changzhi,”A MODWT based algorithm for the identification and removal of jumps/short-term distortions in displacement measurements used for structural health monitoring”,Dec-2021.
Annapurna p Patil,Lakshmi j itagi,ashika cs ,ambika g,mallika ravi,”Design and implementation of an audio fingerprinting system for the identification of audio recordings”,”9th region IEEE conference,2021.
Sungkyun chang,Donmoon lee,”Neural audio fingerprint for high-specific audio retrieval based on contrastive learning”,”IEEE international conference on acoustics”,June-2021.
Jose juan garcia hernandez,Juan jose gomez ricardez,”Hardware architecture for an audio fingerprinting system”,”Computers & electrical engineering,Vol 74,Mar-2019.
Yuexing chen,Jiarun Li,”Recurrent neural networks algorithms and applications”,”2nd international conference IEEE,2021.
NM Alharbi,”Evaluation of sentimental analysis via word embedding and RNN variants for amazon online reviews”,”Mathematical problems in engineering”,2021.
Shirali kadyrov,Cemil turan,altynbek amirzhanov,cemal ozdemir,”Speaker recognition from spectrogram images”,”IEEE international conference”,2021.
Rafizah mohd hanifa,Khalid isa,Shamsul mohamad,”A review on speaker recognition technology and challenges”,”Computer and electrical engineering”,Vol 90,Mar-2021.
Mahdi barhoush,ahmed hallawa,anke schmenik,”Robust automatic speaker identification system using shuffled MFCC features”,”IEEE international conference,2021.
Shabnam farsiani,habib izadkhah,Shahriar Lotfi,”An optimum end to end text independent speaker identification using conventional neural network”,”Computers and electrical engineering”,May 2022.
Vincent roger,jerome farinas,julien pinquier,”Deep neural networks for automatic speech processing a survey from large corpora to limited data”,”EURASIP journal on sudio speech and music processing”,Aug-2022.
Benjamin Lindemann, Timo Müller, Hannes Vietz, Nasser Jazdi, Michael Weyrich,” A survey on long short-term memory networks for time series prediction”,” 14th CIRP Conference on Intelligent Computation in Manufacturing Engineering, CIRP ICME ‘20”,2021.
Downloads
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.