What is it about?

This study elucidated the roles of phonation types and the decoders’ gender in Mandarin emotional speech recognition using the technique of articulatory speech synthesis. Specifically, the phonation types contained breathy, modal, and pressed voices, and the emotional categories included five emotions such as happiness, anger, fear, sadness, and neutrality. It is concluded that pressed voice is closely related to the recognition of anger and breathy voice to the recognition of fear, while the recognition of sadness and happiness does not rely on phonation types to a great degree. Concerning gender differences, females outperformed males in recognizing fear with different phonation types. Females connected modal to pressed voice with neutrality while males linked modal voice to neutrality.

Featured Image

Why is it important?

This is the first study to investigate the roles of phonation types and the decoders’ gender in the recognition of Mandarin emotional speech by means of articulatory speech synthesis. This study provides further evidence that phonation types play a crucial role in the recognition of emotional speech and that gender differences exist. The findings of this study are likely to be helpful to synthesize speech and promote robust automatic recognition of emotional speech. Furthermore, the findings of this study might have implications for diagnosis and intervention among clinical populations with hearing impairment or gender-related psychiatric disorders.

Perspectives

I hope this article can give people a clearer and deeper understanding of the roles of phonation types and decoders' gender in recognizing Mandarin emotional speech. Moreover, I hope this article can help synthesize high-quality emotional speech in Mandarin.

Yajie Chang
Hunan University

Read the Original

This page is a summary of: Roles of Phonation Types and Decoders' Gender in Recognizing Mandarin Emotional Speech, Journal of Speech Language and Hearing Research, October 2023, American Speech-Language-Hearing Association (ASHA),
DOI: 10.1044/2023_jslhr-23-00356.
You can read the full text:

Read

Contributors

The following have contributed to this page