Blood Pressure Estimation from Speech Recordings: Exploring the Role of Voice-over Artists

Main Article Content

Vaishali Rajput
Preeti Mulay
Rajeev Raje

Abstract

Hypertension, a prevalent global health concern, is associated with cardiovascular diseases and significant morbidity and mortality. Accurate and prompt Blood Pressure monitoring is crucial for early detection and successful management. Traditional cuff-based methods can be inconvenient, leading to the exploration of non-invasive and continuous estimation methods. This research aims to bridge the gap between speech processing and health monitoring by investigating the relationship between speech recordings and Blood Pressure estimation. Speech recordings offer promise for non-invasive Blood Pressure estimation due to the potential link between vocal characteristics and physiological responses. In this study, we focus on the role of Voice-over Artists, known for their ability to convey emotions through voice. By exploring the expertise of Voice-over Artists in controlling speech and expressing emotions, we seek valuable insights into the potential correlation between speech characteristics and Blood Pressure. This research sheds light on presenting an innovative and convenient approach to health assessment. By unraveling the specific role of Voice-over Artists in this process, the study lays the foundation for future advancements in healthcare and human-robot interactions. Through the exploration of speech characteristics and emotional expression, this investigation offers valuable insights into the correlation between vocal features and Blood Pressure levels. By leveraging the expertise of Voice-over Artists in conveying emotions through voice, this study enriches our understanding of the intricate relationship between speech recordings and physiological responses, opening new avenues for the integration of voice-related factors in healthcare technologies.

Article Details

How to Cite
Rajput, V. ., Mulay, P. ., & Raje, R. . (2023). Blood Pressure Estimation from Speech Recordings: Exploring the Role of Voice-over Artists. International Journal on Recent and Innovation Trends in Computing and Communication, 11(10s), 152–161. https://doi.org/10.17762/ijritcc.v11i10s.7608
Section
Articles

References

Argha, A., Celler, B. G., & Lovell, N. H. (2021). A Novel Automated Blood Pressure Estimation Algorithm Using Sequences of Korotkoff Sounds. IEEE Journal of Biomedical and Health Informatics, 25(4), 1257–1264. https://doi.org/10.1109/JBHI.2020.3012567

Badshah, A. M., Rahim, N., Ullah, N., Ahmad, J., Muhammad, K., Lee, M. Y., Kwon, S., & Baik, S. W. (2019). Deep features-based speech emotion recognition for smart affective services. Multimedia Tools and Applications, 78(5), 5571–5589. https://doi.org/10.1007/s11042-017-5292-7

Farki, A., Kazemzadeh, R. B., & Noughabi, E. A. (2021). A Novel Clustering-Based Algorithm for Continuous and Non-invasive Cuff-Less Blood Pressure Estimation. In arXiv physics.med-ph]. arXiv. http://arxiv.org/abs/2110.06996

Harfiya, L. N., Chang, C.-C., & Li, Y.-H. (2021). Continuous Blood Pressure Estimation Using Exclusively Photopletysmography by LSTM-Based Signal-to-Signal Translation. Sensors , 21(9). https://doi.org/10.3390/s21092952

Yang, S., Zhang, Y., Cho, S.-Y., Correia, R., & Morgan, S. P. (2021). Non-invasive cuff-less Blood Pressure estimation using a hybrid deep learning model. In Optical and Quantum Electronics (Vol. 53, Issue 2). https://doi.org/10.1007/s11082-020-02667-0

Mottaghi, S., Moradi, M. H., & Moghavvemi, M. (2014). Neuro-fuzzy Indirect Blood Pressure Estimation during Bruce Stress Test.

Guizzo, E., Weyde, T., & Leveson, J. B. (2020). Multi-Time-Scale Convolution for Emotion Recognition from Speech Audio Signals. ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 6489–6493. https://doi.org/10.1109/ICASSP40776.2020.9053727

Mustaqeem, Mustaqeem, Sajjad, M., & Kwon, S. (2020). Clustering-Based Speech Emotion Recognition by Incorporating Learned Features and Deep BiLSTM. In IEEE Access (Vol. 8, pp. 79861–79875). https://doi.org/10.1109/access.2020.2990405

Martínez, G., Howard, N., Abbott, D., Lim, K., Ward, R., & Elgendi, M. (2018). Can Photoplethysmography Replace Arterial Blood Pressure in the Assessment of Blood Pressure? Journal of Clinical Medicine Research, 7(10). https://doi.org/10.3390/jcm7100316

Sakai, M. (2015a). Case study on analysis of vocal frequency to estimate Blood Pressure. 2015 IEEE Congress on Evolutionary Computation, CEC 2015 - Proceedings, 00(c), 2335–2340. https://doi.org/10.1109/CEC.2015.7257173

Sakai, M. (2015b). Modeling the Relationship between Heart Rate and Features of Vocal Frequency. International Journal of Computer Applications, 120(6), 32–37. https://doi.org/10.5120/21233-3986

Ogedegbe, G., & Pickering, T. (2010). Principles and Techniques of Blood Pressure Measurement. Cardiology Clinics, 28(4), 571–586. https://doi.org/10.1016/j.ccl.2010.07.006

Pechprasarn, S., Sukkasem, C., Sasivimolkul, S., & Suvarnaphaet, P. (2019). Machine Learning to identify factors that affect Human Systolic Blood Pressure. BMEiCON 2019 - 12th Biomedical Engineering International Conference. https://doi.org/10.1109/BMEiCON47515.2019.8990356

Ryskaliyev, A., Askaruly, S., & James, A. P. (2016). Speech signal analysis for the estimation of heart rates under different emotional states. 2016 International Conference on Advances in Computing, Communications and Informatics, ICACCI 2016, 1160–1165. https://doi.org/10.1109/ICACCI.2016.7732201

Krishna Kishore, K. V., & Krishna Satish, P. (2013). Emotion recognition in speech using MFCC and wavelet features. Proceedings of the 2013 3rd IEEE International Advance Computing Conference, IACC 2013, 842–847. https://doi.org/10.1109/IAdCC.2013.6514336

Kachuee, M., Kiani, M. M., Mohammadzade, H., & Shabany, M. (2017). Cuffless Blood Pressure Estimation Algorithms for Continuous Health-Care Monitoring. IEEE Transactions on Biomedical Engineering, 64(4), 859–869. https://doi.org/10.1109/TBME.2016.2580904

El Ayadi, M., Kamel, M. S., & Karray, F. (2011). Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognition, 44(3), 572–587. https://doi.org/10.1016/j.patcog.2010.09.020

Aich, S., Kim, H. -., Younga, K., Hui, K. L., Al-Absi, A. A., & Sain, M. (2019). A supervised machine learning approach using different feature selection techniques on voice datasets for prediction of parkinson's disease. Paper presented at the International Conference on Advanced Communication Technology, ICACT, 2019-February 1116-1121. doi:10.23919/ICACT.2019.8701961 Retrieved from www.scopus.com

Mustaqeem, Mustaqeem, & Kwon, S. (2021). MLT-DNet: Speech emotion recognition using 1D dilated CNN based on multi-learning trick approach. In Expert Systems with Applications (Vol. 167, p. 114177). https://doi.org/10.1016/j.eswa.2020.114177

Nie, W., Ren, M., Nie, J., & Zhao, S. (2021). C-GCN: Correlation Based Graph Convolutional Network for Audio-Video Emotion Recognition. IEEE Transactions on Multimedia, 23, 3793–3804. https://doi.org/10.1109/TMM.2020.3032037

Mustaqeem, & Kwon, S. (2019). A CNN-Assisted Enhanced Audio Signal Processing for Speech Emotion Recognition. Sensors , 20(1). https://doi.org/10.3390/s20010183

Fahad Nasser Alhazmi. (2023). Self-Efficacy and Personal Innovation: Conceptual Model Effects on Patients’ Perceptions of PHR Use in Saudi Arabia. International Journal of Intelligent Systems and Applications in Engineering, 11(4s), 369–376. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/2676

Alkan, A., & Günay, M. (2012). Identification of EMG signals using discriminant analysis and SVM classifier. Expert Systems with Applications, 39(1), 44–47. https://doi.org/10.1016/j.eswa.2011.06.043

Pan, Y., Shen, P., & Shen, L. (2012). Speech emotion recognition using support vector machine. International Journal of Smart Home, 6(2), 101–108. https://doi.org/10.1109/kst.2013.6512793

Mustaqeem, Mustaqeem, & Kwon, S. (2020). CLSTM: Deep Feature-Based Speech Emotion Recognition Using the Hierarchical ConvLSTM Network. In Mathematics (Vol. 8, Issue 12, p. 2133). https://doi.org/10.3390/math8122133

Song, K., Park, T.-J., & Chang, J.-H. (2021). Novel Data Augmentation Employing Multivariate Gaussian Distribution for Neural Network-Based Blood Pressure Estimation. NATO Advanced Science Institutes Series E: Applied Sciences, 11(9), 3923. https://doi.org/10.3390/app11093923

Y., A., A., K., A., Y., & Pandya, N. (2017). Survey paper on Different Speech Recognition Algorithm: Challenges and Techniques. International Journal on Speech Communications, 175(1), 31–36. https://doi.org/10.5120/ijca2017915472

Shakti P.Rath(2017). Scalable algorithms for unsupervised clustering of acoustic data for speech recognition. Computer Speech & Language. Volume 46, November 2017, Pages 233-248. https://doi.org/10.1016/j.csl.2017.06.001

T. B. Amin, P. Marziliano and J. S. German, "Nine Voices, One Artist: Linguistic and Acoustic Analysis," 2012 IEEE International Conference on Multimedia and Expo, Melbourne, VIC, Australia, 2012, pp. 450-454, doi: 10.1109/ICME.2012.142.

N. Obin and A. Roebel, "Similarity Search of Acted Voices for Automatic Voice Casting," in IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 24, no. 9, pp. 1642-1651, Sept. 2016, doi: 10.1109/TASLP.2016.2580302.

Phalguni Phatak, Sonia Shaikh, Neha Jamdhade & Pallavi Sovani Kelkar (2021) Do Voice-Over Artists Convey Emotion Better Than Untrained Voice Users?, Voice and Speech Review, 15:3, 315-329, DOI: 10.1080/23268263.2021.1882751

Duan, KB., Rajapakse, J.C., Nguyen, M.N. (2007). One-Versus-One and One-Versus-All Multiclass SVM-RFE for Gene Selection in Cancer Classification. In: Marchiori, E., Moore, J.H., Rajapakse, J.C. (eds) Evolutionary Computation,Machine Learning and Data Mining in Bioinformatics. EvoBIO 2007. Lecture Notes in Computer Science, vol 4447. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-71783-6_5

Dhabliya, M. D. . (2021). Cloud Computing Security Optimization via Algorithm Implementation. International Journal of New Practices in Management and Engineering, 10(01), 22–24. https://doi.org/10.17762/ijnpme.v10i01.99

Hong Feng and Xunbing Shen. 2022. A random forest algorithm-based emotion recognition model for eye features. In Proceedings of the 3rd International Symposium on Artificial Intelligence for Medicine Sciences (ISAIMS '22). Association for Computing Machinery, New York, NY, USA, 148–152. https://doi.org/10.1145/3570773.3570851

Hariharan Muthusamy, Kemal Polat, Sazali Yaacob, "Improved Emotion Recognition Using Gaussian Mixture Model and Extreme Learning Machine in Speech and Glottal Signals", Mathematical Problems in Engineering, vol. 2015, Article ID 394083, 13 pages, 2015. https://doi.org/10.1155/2015/394083

Xing, W., & Bei, Y. (2020). Medical Health Big Data Classification Based on KNN Classification Algorithm. IEEE Access, 8, 28808–28819. https://doi.org/10.1109/access.2019.2955754

An, Y., Sun, S., & Wang, S. (2017). Naive Bayes classifiers for music emotion classification based on lyrics. https://doi.org/10.1109/icis.2017.7960070

Mark White, Thomas Wood, Carlos Rodríguez, Pekka Koskinen, Jónsson Ólafur. Exploring Natural Language Processing in Educational Applications. Kuwait Journal of Machine Learning, 2(1). Retrieved from http://kuwaitjournals.com/index.php/kjml/article/view/168

Liu, Z., Wu, M., Cao, W., Mao, J., Xu, J., & Tan, G. (2018). Speech emotion recognition based on feature selection and extreme learning machine decision tree. Neurocomputing, 273, 271–280. https://doi.org/10.1016/j.neucom.2017.07.050

Zhang, W. ., & Chung Ee, J. Y. . (2023). AnIntelligent Knowledge Graph-Based Directional Data Clustering and Feature Selection for Improved Education. International Journal on Recent and Innovation Trends in Computing and Communication, 11(6s), 22–33. https://doi.org/10.17762/ijritcc.v11i6s.6807

38. Velunachiyar, S., & Sivakumar, K. (2023). Some Clustering Methods, Algorithms and their Applications. International Journal on Recent and Innovation Trends in Computing and Communication, 11(6s),401–410. https://doi.org/10.17762/ijritcc.v11i6s.6946