Comparing supervised and self-supervised embedding for ExVo Multi-Task learning track
Type of publication: | Conference paper |
Citation: | Purohit_ICML_2022 |
Publication status: | Accepted |
Booktitle: | Proceedings of the ICML Expressive Vocalizations Workshop held in conjunction with the 39th International Conference on Machine Learning |
Year: | 2022 |
Location: | Maryland, USA |
Abstract: | The ICML Expressive Vocalizations (ExVo) Multi-task challenge 2022, focuses on understanding the emotional facets of the non-linguistic vocalizations (vocal bursts (VB)). The objective of this challenge is to predict emotional intensities for VB, being a multi-task challenge it also requires to predict speakers' age and native-country. For this challenge we study and compare two distinct embedding spaces namely, self-supervised learning (SSL) based embeddings and task-specific supervised learning based embeddings. Towards that, we investigate feature representations obtained from several pre-trained SSL neural networks and task-specific supervised classification neural networks. Our studies show that the best performance is obtained with an hybrid approach, where predictions derived via both SSL and task-specific supervised learning are used. Our best system on test-set surpass the ComPARE baseline (harmonic-mean of all sub-task scores i.e., S_MLT) by a relative 13% margin. |
Keywords: | Emotion Recognition, Expressive Vocalizations, Multi-task learning, Self-supervised embedding |
Projects |
EMIL |
Authors | |
Added by: | [UNK] |
Total mark: | 0 |
Attachments
|
|
Notes
|
|
|