What is it about?

This paper explains our solutions for the Stuttering and the Vocalisation tasks within the ACM Multimedia 2022 Computational Paralinguistics ChallengE. The goal of this competition is to develop Artificial Intelligence models that can recognize different aspects (stutterings and emotions) of the speaker using only their recorded voice.

Featured Image

Why is it important?

Our system achieved excellent results in both competitions and showcases how we can recognize different kinds of stutterings (e.g. word/phrase repetition, prolongation, sound repetition, etc.) and emotions in non-verbal vocal expressions (such as laughter, cries, moans, and screams). In both challenges, our solution achieved the best results, demonstrating their effectiveness.

Perspectives

Participating in these competitions is quite important for me. On the one hand, it poses new challenges for me and often makes me think "outside of the box", to come up with novel solutions. At the same time, it allows us to compete with other groups in a controlled setup, using the same data and under the same limitations.

Tamas Grosz
Aalto-yliopisto

Read the Original

This page is a summary of: Wav2vec2-based Paralinguistic Systems to Recognise Vocalised Emotions and Stuttering, October 2022, ACM (Association for Computing Machinery),
DOI: 10.1145/3503161.3551572.
You can read the full text:

Read

Contributors

The following have contributed to this page