Journal of Machine and Computing


An Efficient Voice Authentication System using Enhanced Inceptionv3 Algorithm



Journal of Machine and Computing

Received On : 04 February 2023

Revised On : 18 May 2023

Accepted On : 18 June 2023

Published On : 05 October 2023

Volume 03, Issue 04

Pages : 379-393


Abstract


Automatic voice authentication based on deep learning is a promising technology that has received much attention from academia and industry. It has proven to be effective in a variety of applications, including biometric access control systems. Using biometric data in such systems is difficult, particularly in a centralized setting. It introduces numerous risks, such as information disclosure, unreliability, security, privacy, etc. Voice authentication systems are becoming increasingly important in solving these issues. This is especially true if the device relies on voice commands from the user. This work investigates the development of a text-independent voice authentication system. The spatial features of the voiceprint (corresponding to the speech spectrum) are present in the speech signal as a result of the spectrogram, and the weighted wavelet packet cepstral coefficients (W-WPCC) are effective for spatial feature extraction (corresponding to the speech spectrum). W-WPCC characteristics are calculated by combining sub-band energies with sub-band spectral centroids using a weighting scheme to generate noise-resistant acoustic characteristics. In addition, this work proposes an enhanced inception v3 model for voice authentication. The proposed InceptionV3 system extracts feature from input data from the convolutional and pooling layers. By employing fewer parameters, this architecture reduces the complexity of the convolution process while increasing learning speed. Following model training, the enhanced Inception v3 model classifies audio samples as authenticated or not based on extracted features. Experiments were carried out on the speech of five English speakers whose voices were collected from YouTube. The results reveal that the suggested improved method, based on enhanced Inception v3 and trained on speech spectrogram pictures, outperforms the existing methods. The approach generates tests with an average categorization accuracy of 99%. Compared to the performance of these network models on the given dataset, the proposed enhanced Inception v3 network model achieves the best results regarding model training time, recognition accuracy, and stability.


Keywords


Voice Authentication, Short-Time Fourier Transform, Weighted Wavelet Packet Cepstral Coefficient, Inception V3.


  1. H. Park and T. Kim, “User Authentication Method via Speaker Recognition and Speech Synthesis Detection,” Security and Communication Networks, vol. 2022, pp. 1–10, Jan. 2022, doi: 10.1155/2022/5755785.
  2. S. K. Wong and S. M. Yiu, “Location Spoofing Attack Detection with Pre-Installed Sensors in Mobile Devices,” Journal of Wireless Mobile Networks, Ubiquitous Computing, and Dependable Applications (JoWUA), vol. 11, no. 4, pp. 16–30, Dec. 2020, doi: 10.22667/JOWUA.2020.12.31.016.
  3. A. S. Kitana, T. Issa, and W. G. Isaac, “Towards an Epidemic SMS-based Cellular Botnet,” Journal of Internet Services and Information Security (JISIS), vol. 10, no. 4, pp. 38–58, Nov. 2020, doi: 10.22667/JISIS.2020.11.30.038.
  4. G. S. Kasturi, A. Jain, and J. D. Singh, “Detection and Classification of Radio Frequency Jamming Attacks using Machine learning,” Journal of Wireless Mobile Networks, Ubiquitous Computing, and Dependable Applications (JoWUA), vol. 11, no. 4, pp. 49–62, Dec. 2020, doi: 10.22667/JOWUA.2020.12.31.049.
  5. A. L. Marra, F. Martinelli, F. Mercaldo, A. Saracino, and M. Sheikhalishahi, “A Distributed Framework for Collaborative and Dynamic Analysis of Android Malware,” Journal of Wireless Mobile Networks, Ubiquitous Computing, and Dependable Applications (JoWUA), vol. 11, no. 3, pp. 1–28, Sep. 2020, doi: 10.22667/JOWUA.2020.09.30.001.
  6. D. Berbecaru, A. Lioy, and C. Cameroni, “Supporting Authorize-then-Authenticate for Wi-Fi access based on an Electronic Identity Infrastructure,” Journal of Wireless Mobile Networks, Ubiquitous Computing, and Dependable Applications (JoWUA), vol. 11, no. 2, pp. 34– 54, June. 2020, doi: 10.22667/JOWUA.2020.06.30.034.
  7. S. H. K. Wong and S. M. Yiu, “Identification of device motion status via Bluetooth discovery,” Journal of Internet Services a nd Information Security (JISIS), vol. 10, no. 4, pp. 59–69, Nov. 2020, doi: 10.22667/JISIS.2020.11.30.059.
  8. J. A. Unar, W. C. Seng, and A. Abbasi, “A review of biometric technology along with trends and prospects,” Pattern Recognition, vol. 47, no. 8, pp. 2673–2688, Aug. 2014, doi: 10.1016/j.patcog.2014.01.016.
  9. D. A. Reynolds, T. F. Quatieri, and R. B. Dunn, “Speaker Verification Using Adapted Gaussian Mixture Models,” Digital Signal Processing, vol. 10, no. 1–3, pp. 19–41, Jan. 2000, doi: 10.1006/dspr.1999.0361.
  10. D. A. Reynolds and R. C. Rose, “Robust text-independent speaker identification using Gaussian mixture speaker models,” IEEE Transactions on Speech and Audio Processing, vol. 3, no. 1, pp. 72–83, 1995, doi: 10.1109/89.365379.
  11. N. H. Tandel, H. B. Prajapati, and V. K. Dabhi, “Voice Recognition and Voice Comparison using Machine Learning Techniques: A Survey,” 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS), Mar. 2020, doi: 10.1109/icaccs48705.2020.9074184.
  12. H. Y. Khdier, W. M. Jasim, and S. A. Aliesawi, “Deep Learning Algorithms based Voiceprint Recognition System in Noisy Environment,” Journal of Physics: Conference Series, vol. 1804, no. 1, p. 012042, Feb. 2021, doi: 10.1088/1742-6596/1804/1/012042.
  13. K. Aizat, O. Mohamed, M. Orken, A. Ainur, and B. Zhumazhanov, “Identification and authentication of user voice using DNN features and i- vector,” Cogent Engineering, vol. 7, no. 1, p. 1751557, Jan. 2020, doi: 10.1080/23311916.2020.1751557.
  14. T. Zeng, “Deep Learning in Automatic Speech Recognition (ASR): A Review,” Proceedings of the 2022 7th International Conference on Modern Management and Education Technology (MMET 2022), pp. 173–179, Dec. 2022, doi: 10.2991/978-2-494069-51-0_23.
  15. A. Alsobhani, H. M. A. ALabboodi, and H. Mahdi, “Speech Recognition using Convolution Deep Neural Networks,” Journal of Physics: Conference Series, vol. 1973, no. 1, p. 012166, Aug. 2021, doi: 10.1088/1742-6596/1973/1/012166.
  16. R. Zheng, Y. Fang, and J. Dong, “Voice Print Recognition Check-in System Based on Resnet,” Highlights in Science, Engineering and Technology, vol. 16, pp. 98–108, Nov. 2022, doi: 10.54097/hset.v16i.2473.
  17. F. Ye and J. Yang, “A Deep Neural Network Model for Speaker Identification,” Applied Sciences, vol. 11, no. 8, p. 3603, Apr. 2021, doi: 10.3390/app11083603.
  18. Bella, J. Hendryli, and D. E. Herwindiati, “Voice Authentication Model for One-time Password Using Deep Learning Models,” Proceedings of the 2020 2nd International Conference on Big Data Engineering and Technology, Jan. 2020, doi: 10.1145/3378904.3378908.
  19. T. Muruganantham, N. R. NAGARAJAN, and R. Balamurugan, "Biometric Of Speaker Authentication Using CNN,". 13. 1417-1423.
  20. S. Duraibi, W. Alhamdani, and F. T. Sheldon, “Voice Feature Learning using Convolutional Neural Networks Designed to Avoid Replay Attacks,” 2020 IEEE Symposium Series on Computational Intelligence (SSCI), Dec. 2020, doi: 10.1109/ssci47803.2020.9308489.
  21. A. Shafik et al., “Speaker identification based on Radon transform and CNNs in the presence of different types of interference for Robotic Applications,” Applied Acoustics, vol. 177, p. 107665, Jun. 2021, doi: 10.1016/j.apacoust.2020.107665.
  22. A. B. Abdusalomov, F. Safarov, M. Rakhimov, B. Turaev, and T. K. Whangbo, “Improved Feature Parameter Extraction from Speech Signals Using Machine Learning Algorithm,” Sensors, vol. 22, no. 21, p. 8122, Oct. 2022, doi: 10.3390/s22218122.
  23. W. Jia, L. Dongmei, "A review of deep learning applications in speech recognition," Computer Knowledge and Technology, 13(16): 191-197, 2020.
  24. M. Han, T. Roubing, Z. Yi, et al. "Survey on Speech Recognition," Computer Systems & Applications, 31(1):1−10, 2022.
  25. M. Wollmer, F. Eyben, B. Schuller, and G. Rigoll, “A multi-stream ASR framework for BLSTM modeling of conversational speech,” 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), May 2011, doi: 10.1109/icassp.2011.5947444.
  26. Wen-kai Lu and Qiang Zhang, “Deconvolutive Short-Time Fourier Transform Spectrogram,” IEEE Signal Processing Letters, vol. 16, no. 7, pp. 576–579, Jul. 2009, doi: 10.1109/lsp.2009.2020887.
  27. Y. Huang, K. Tian, A. Wu, and G. Zhang, “Feature fusion methods research based on deep belief networks for speech emotion recognition under noise condition,” Journal of Ambient Intelligence and Humanized Computing, vol. 10, no. 5, pp. 1787–1798, Dec. 2017, doi: 10.1007/s12652-017- 0644-8.
  28. J. Cao, M. Yan, Y. Jia, X. Tian, and Z. Zhang, “Application of a modified Inception-v3 model in the dynasty-based classification of ancient murals,” EURASIP Journal on Advances in Signal Processing, vol. 2021, no. 1, Jul. 2021, doi: 10.1186/s13634-021-00740-8.
  29. Q. Zou, Y. Cao, Q. Li, C. Huang, and S. Wang, “Chronological classification of ancient paintings using appearance and shape features,” Pattern Recognition Letters, vol. 49, pp. 146–154, Nov. 2014, doi: 10.1016/j.patrec.2014.07.002.
  30. S. Raj, P. Prakasam, and S. Gupta, “Audio signal quality enhancement using multi-layered convolutional neural network based auto encoder– decoder,” International Journal of Speech Technology, vol. 24, no. 2, pp. 425–437, Jan. 2021, doi: 10.1007/s10772-021-09809-z.

Acknowledgements


The author(s) received no financial support for the research, authorship, and/or publication of this article.


Funding


No funding was received to assist with the preparation of this manuscript.


Ethics declarations


Conflict of interest

The authors have no conflicts of interest to declare that are relevant to the content of this article.


Availability of data and materials


No data available for above study.


Author information


Contributions

All authors have equal contribution in the paper and all authors have read and agreed to the published version of the manuscript.


Corresponding author


Rights and permissions


Open Access This article is licensed under a Creative Commons Attribution NoDerivs is a more restrictive license. It allows you to redistribute the material commercially or non-commercially but the user cannot make any changes whatsoever to the original, i.e. no derivatives of the original work. To view a copy of this license, visit https://creativecommons.org/licenses/by-nc-nd/4.0/


Cite this article


Kaladharan N and Arunkumar R, “An Efficient Voice Authentication System using Enhanced Inceptionv3 Algorithm”, Journal of Machine and Computing, vol.3, no.4, pp. 379-393, October 2023. doi: 10.53759/7669/jmc202303032.


Copyright


© 2023 Kaladharan N and Arunkumar R. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.