Development of Polymer-Based Sensors for Speech Emotion Recognition

Open Access

Year : 2024 | Volume :12 | Special Issue : 05 | Page : 268-274
By

Princy Tyagi,

  1. Assistant Professor, Computer Science & Engineering, Swami Rama Himalayan University, Dehradun, Uttarakhand, India

Abstract

Traditional SER research often utilizes microphones with polymer components like Diaphragms and Membranes. Within some microphone designs, polymer membranes which plays a crucial role in converting sound pressure into electrical signals. The paper highlights the application (speech emotion recognition) and have tried to find polymer-based sensors. This work further delves deeper, investigating the performance of the CatBoost algorithm for emotion recognition in voice assistants designed for Indian languages. The research employs a dataset of labelled audio recordings encompassing various emotions in different Indian languages. Mel-Frequency Cepstral Coefficients (MFCC) and pitch features are extracted from the speech data. The CatBoost algorithm is then utilized for classification and compared to other commonly used algorithms like CNN, LSTM, XGBoost, and LightGBM. Our findings demonstrate that CatBoost achieves superior accuracy in emotion recognition compared to the other evaluated algorithms, particularly when using MFCC and pitch features. This highlights the potential of CatBoost for developing robust and efficient SER systems tailored for Indian language voice assistants. This research offers a novel perspective by bridging the gap between machine learning for SER and its potential application in voice assistants designed for specific languages. By leveraging CatBoost’s capabilities, future voice assistants could better understand and respond to the emotional context of user interactions, potentially enhancing the user experience for Indian language speakers.

Keywords: Polymer components, Pattern recognition, Speech Emotion Recognition, Voice Analysis, Machine learning Algorithm.

[This article belongs to Special Issue under section in Journal of Polymer and Composites (jopc)]

How to cite this article:
Princy Tyagi. Development of Polymer-Based Sensors for Speech Emotion Recognition. Journal of Polymer and Composites. 2024; 12(05):268-274.
How to cite this URL:
Princy Tyagi. Development of Polymer-Based Sensors for Speech Emotion Recognition. Journal of Polymer and Composites. 2024; 12(05):268-274. Available from: https://journals.stmjournals.com/jopc/article=2024/view=176877


Full Text PDF

Browse Figures

References

  1. Sharma U, Maheshkar S, Mishra AN. Study of robust feature extraction techniques for speech recognition system. In2015 International conference on futuristic trends on computational analysis and knowledge management (ABLAZE) 2015 Feb 25 (pp. 654-658). IEEE.
  2. Gupta H, Gupta D. LPC and LPCC method of feature extraction in Speech Recognition System. In 2016 6th international conference-cloud system and big data engineering (confluence) 2016 Jan 14 (pp. 498-502). IEEE.
  3. Chadha AN, Zaveri MA, Sarvaiya JN. Optimal feature extraction and selection techniques for speech processing: A review. In 2016 International Conference on Communication and Signal Processing (ICCSP) 2016 Apr 6 (pp. 1669-1673). IEEE.
  4. Letaifa LB, Torres MI, Justo R. Adding dimensional features for emotion recognition on speech. In 2020 5th international conference on advanced technologies for signal and image processing (ATSIP) 2020 Sep 2 (pp. 1-6). IEEE.
  5. Mehra P, Verma SK. Comparing Classifiers for Recognizing the Emotions by extracting the Spectral Features of Speech Using Machine Learning. In 2023 International Conference on Device Intelligence, Computing and Communication Technologies,(DICCT) 2023 Mar 17 (pp. 387-391). IEEE.
  6. Lakomkin E, Zamani MA, Weber C, Magg S, Wermter S. Incorporating end-to-end speech recognition models for sentiment analysis. In 2019 International Conference on Robotics and Automation (ICRA) 2019 May 20 (pp. 7976-7982). IEEE.
  7. Sajjad M, Kwon S. Clustering-based speech emotion recognition by incorporating learned features and deep BiLSTM. IEEE access. 2020 Apr 27;8:79861-75.
  8. Garg K, Jain G. A comparative study of noise reduction techniques for automatic speech recognition systems. In 2016 International Conference on Advances in Computing, Communications and Informatics (ICACCI) 2016 Sep 21 (pp. 2098-2103). IEEE.
  9. Pardede HF. On noise robust feature for speech recognition based on power function family. In 2015 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS) 2015 Nov 9 (pp. 386-390). IEEE.
  10. Alim SA, Rashid NK. Some commonly used speech feature extraction algorithms. London, UK:: IntechOpen; 2018 Dec 12.
  11. Gill AS. A review on feature extraction techniques for speech processing. International Journal Of Engineering and Computer Science. 2016 Oct;5(10):18551-6.
  12. Hermansky H. Perceptual linear predictive (PLP) analysis of speech. the Journal of the Acoustical Society of America. 1990 Apr 1;87(4):1738-52.
  13. Luo Y, Fu Q, Xie J, Qin Y, Wu G, Liu J, Jiang F, Cao Y, Ding X. EEG-based emotion classification using spiking neural networks. IEEE Access. 2020 Mar 4;8:46007-16.
  14. Amini MM, Matrouf D. Data augmentation versus noise compensation for x-vector speaker recognition systems in noisy environments. In 2020 28th European Signal Processing Conference (EUSIPCO) 2021 Jan 18 (pp. 1-5). IEEE.
  15. Wu J, Hua Y, Yang S, Qin H, Qin H. Speech enhancement using generative adversarial network by distilling knowledge from statistical method. Applied Sciences. 2019 Aug 18;9(16):3396.

Special Issue Open Access Original Research
Volume 12
Special Issue 05
Received 01/03/2024
Accepted 16/07/2024
Published 30/07/2024

Check Our other Platform for Workshops in the field of AI, Biotechnology & Nanotechnology.
Check Out Platform for Webinars in the field of AI, Biotech. & Nanotech.