Mostrar el registro sencillo del ítem

dc.contributor.authorNosek, Tijana V.
dc.contributor.authorSuzić, Siniša B.
dc.contributor.authorPekar, Darko J.
dc.contributor.authorObradović, Radovan J.
dc.contributor.authorSečujski, Milan S.
dc.contributor.authorDelić, Vlado D.
dc.date2021-12
dc.date.accessioned2022-05-11T09:49:49Z
dc.date.available2022-05-11T09:49:49Z
dc.identifier.issn1989-1660
dc.identifier.urihttps://reunir.unir.net/handle/123456789/13070
dc.description.abstractThe paper presents a novel architecture and method for speech synthesis in multiple languages, in voices of multiple speakers and in multiple speaking styles, even in cases when speech from a particular speaker in the target language was not present in the training data. The method is based on the application of neural network embedding to combinations of speaker and style IDs, but also to phones in particular phonetic contexts, without any prior linguistic knowledge on their phonetic properties. This enables the network not only to efficiently capture similarities and differences between speakers and speaking styles, but to establish appropriate relationships between phones belonging to different languages, and ultimately to produce synthetic speech in the voice of a certain speaker in a language that he/she has never spoken. The validity of the proposed approach has been confirmed through experiments with models trained on speech corpora of American English and Mexican Spanish. It has also been shown that the proposed approach supports the use of neural vocoders, i.e. that they are able to produce synthesized speech of good quality even in languages that they were not trained on.es_ES
dc.language.isoenges_ES
dc.publisherInternational Journal of Interactive Multimedia and Artificial Intelligence (IJIMAI)es_ES
dc.relation.ispartofseries;vol. 7, nº 2
dc.relation.urihttps://www.ijimai.org/journal/bibcite/reference/3049es_ES
dc.rightsopenAccesses_ES
dc.subjectcross-linguales_ES
dc.subjectartificial neural networkses_ES
dc.subjectspeech synthesises_ES
dc.subjectvocoderes_ES
dc.subjectIJIMAIes_ES
dc.titleCross-Lingual Neural Network Speech Synthesis Based on Multiple Embeddingses_ES
dc.typearticlees_ES
reunir.tag~IJIMAIes_ES
dc.identifier.doihttps://doi.org/10.9781/ijimai.2021.11.005


Ficheros en el ítem

Thumbnail

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem