Novel Task-Based Unification and Adaptation (TUA) Transfer Learning Approach for Bilingual Emotional Speech Data

0Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

Modern developments in machine learning methodology have produced effective approaches to speech emotion recognition. The field of data mining is widely employed in numerous situations where it is possible to predict future outcomes by using the input sequence from previous training data. Since the input feature space and data distribution are the same for both training and testing data in conventional machine learning approaches, they are drawn from the same pool. However, because so many applications require a difference in the distribution of training and testing data, the gathering of training data is becoming more and more expensive. High performance learners that have been trained using similar, already-existing data are needed in these situations. To increase a model’s capacity for learning, transfer learning involves transferring knowledge from one domain to another related domain. To address this scenario, we have extracted ten multi-dimensional features from speech signals using OpenSmile and a transfer learning method to classify the features of various datasets. In this paper, we emphasize the importance of a novel transfer learning system called Task-based Unification and Adaptation (TUA), which bridges the disparity between extensive upstream training and downstream customization. We take advantage of the two components of the TUA, task-challenging unification and task-specific adaptation. Our algorithm is studied using the following speech datasets: the Arabic Emirati-accented speech dataset (ESD), the English Speech Under Simulated and Actual Stress (SUSAS) dataset and the Ryerson Audio-Visual Database of Emotional Speech and Song dataset (RAVDESS). Using the multidimensional features and transfer learning method on the given datasets, we were able to achieve an average speech emotion recognition rate of 91.2% on the ESD, 84.7% on the RAVDESS and 88.5% on the SUSAS datasets, respectively.

References Powered by Scopus

Deep residual learning for image recognition

163438Citations
N/AReaders
Get full text

Long Short-Term Memory

73181Citations
N/AReaders
Get full text

Feature pyramid networks for object detection

18677Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Shahin, I., Nassif, A. B., Thomas, R., & Hamsa, S. (2023). Novel Task-Based Unification and Adaptation (TUA) Transfer Learning Approach for Bilingual Emotional Speech Data. Information (Switzerland), 14(4). https://doi.org/10.3390/info14040236

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 2

50%

Lecturer / Post doc 1

25%

Researcher 1

25%

Readers' Discipline

Tooltip

Computer Science 4

80%

Engineering 1

20%

Article Metrics

Tooltip
Mentions
Blog Mentions: 1
News Mentions: 1

Save time finding and organizing research with Mendeley

Sign up for free