Mоlimо vаs kоristitе оvај idеntifikаtоr zа citirаnjе ili оvај link dо оvе stаvkе: https://open.uns.ac.rs/handle/123456789/2043
Pоljе DC-аVrеdnоstЈеzik
dc.contributor.authorSuzić, Sinišaen
dc.contributor.authorDelić, Tijanaen
dc.contributor.authorOstrogonac S.en
dc.contributor.authorĐuričić , Jasnaen
dc.contributor.authorPekar, Darkoen
dc.date.accessioned2019-09-23T10:19:14Z-
dc.date.available2019-09-23T10:19:14Z-
dc.date.issued2018-01-01en
dc.identifier.issn20789181en
dc.identifier.urihttps://open.uns.ac.rs/handle/123456789/2043-
dc.description.abstract© 2018 St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences. All rights reserved. Modern text-to-speech systems generally achieve good intelligibility. The one of the main drawbacks of these systems is the lack of expressiveness in comparison to natural human speech. It is very unpleasant when automated system conveys positive and negative message in completely the same way. The introduction of parametric methods in speech synthesis gave possibility to easily change speaker characteristics and speaking styles. In this paper a simple method for incorporating styles into synthesized speech by using style codes is presented. The proposed method requires just a couple of minutes of target style and moderate amount of neutral speech. It is successfully applied to both hidden Markov models and deep neural networks-based synthesis, giving style code as additional input to the model. Listening tests confirmed that better style expressiveness is achieved by deep neural networks synthesis compared to hidden Markov model synthesis. It is also proved that quality of speech synthesized by deep neural networks in a certain style is comparable with the speech synthesized in neutral style, although the neutral-speech-database is about 10 times bigger. DNN based TTS with style codes are further investigated by comparing the quality of speech produced by single-style modeling and multi-style modeling systems. Objective and subjective measures confirmed that there is no significant difference between these two approaches.en
dc.relation.ispartofSPIIRAS Proceedingsen
dc.titleStyle-code method for multi-style parametric text-to-speech syntesisen
dc.typeJournal/Magazine Articleen
dc.identifier.doi10.15622/sp.60.8en
dc.identifier.scopus2-s2.0-85057534493en
dc.identifier.urlhttps://api.elsevier.com/content/abstract/scopus_id/85057534493en
dc.relation.lastpage240en
dc.relation.firstpage216en
dc.relation.issue60en
dc.relation.volume5en
item.fulltextNo Fulltext-
item.grantfulltextnone-
crisitem.author.deptFakultet tehničkih nauka, Departman za energetiku, elektroniku i telekomunikacije-
crisitem.author.parentorgFakultet tehničkih nauka-
Nаlаzi sе u kоlеkciјаmа:FTN Publikacije/Publications
Prikаzаti јеdnоstаvаn zаpis stаvki

SCOPUSTM   
Nаvоđеnjа

2
prоvеrеnо 20.05.2023.

Prеglеd/i stаnicа

25
Prоtеklа nеdеljа
10
Prоtеkli mеsеc
0
prоvеrеnо 10.05.2024.

Google ScholarTM

Prоvеritе

Аlt mеtrikа


Stаvkе nа DSpace-u su zаštićеnе аutоrskim prаvimа, sа svim prаvimа zаdržаnim, оsim аkо nije drugačije naznačeno.