Please use this identifier to cite or link to this item: https://repositori.mypolycc.edu.my/jspui/handle/123456789/6805
Full metadata record
DC FieldValueLanguage
dc.contributor.authorDeng, Guangfeng-
dc.date.accessioned2025-10-13T04:04:38Z-
dc.date.available2025-10-13T04:04:38Z-
dc.date.issued2025-04-30-
dc.identifier.issn2327-5227-
dc.identifier.issn2327-5219-
dc.identifier.otherdoi.org/10.4236/jcc.2025.134020-
dc.identifier.urihttps://repositori.mypolycc.edu.my/jspui/handle/123456789/6805-
dc.description.abstractTelling a story requires various emotional ups and downs as well as pauses. Preparing a parallel corpus for emotional voice conversion is often costly and impractical. Developing high-quality non-parallel methods poses a significant challenge. Although non-parallel methods have been shown to enable emotional voice conversion, its capability for Chinese storytelling has not been clarified. Additionally, the storytelling results of emotional voice conversion have not been validated within a 3-12-year-old children. This study proposes a two-stage Chinese Storytelling Style Speech Generation System (SSPGS) composed of a text-to-speech system and an emotional voice conversion module. The SSPGS requires no parallel utterances, transcriptions, or time alignment procedures for speech generator training and requires only several minutes of training examples to generate reasonably realistic sounding speech. A small corpus neutral speech model is constructed on the text-to-speech system in the first stage, which is based on the speech synthesis system using a Hidden Markov Model (HMM). In the second stage, the emotional voice con version module based on Cycle-Consistent generative adversarial networks (CycleGAN) is built. It enables the neutral speech generated by the text-to speech system in the first stage to be transformed into the happiness, anger, and sadness necessary for storytelling tone using the timbre (spectrum), pitch (fundamental frequency F0), and rhythm (speech rate) of neutral speech. The validity of SSPGS is verified in two ways. First, a 5-point Mean Opinion Score (MOS) was performed for young children’s parents. The results demonstrated that compared with general speech synthesizers, such as Google, the system generated more natural and genuine sound, that was more preferrable to the target audience. After that, the kids underwent a story immersion evaluation. Analysis of the degree of engagement, liking, and empathy in listening to the story revealed no statistically significant difference between real-person dubbing and emotional speech synthesis dubbing. As a result, it has been initially confirmed that SSPGS might be added to the story robot product in the future.ms_IN
dc.language.isoenms_IN
dc.publisherScientific Research Publishing Inc.ms_IN
dc.relation.ispartofseriesJournal of Computer and Communications;2025, 13(4), 324-346-
dc.subjectStorytelling style speech generation systemms_IN
dc.subjectEmotional voice conversion modulems_IN
dc.subjectCycle-consistent generative adversarial networksms_IN
dc.subjectText-to-speech systemms_IN
dc.subjectMean opinion scorems_IN
dc.subjectImmersion measurementms_IN
dc.titleSTORYTELLING STYLE SPEECH GENERATION SYSTEM: EMOTIONAL VOICE CONVERSION MODULE BASED ON CYCLE-CONSISTENT GENERATIVE ADVERSARIAL NETWORKSms_IN
dc.typeArticlems_IN
Appears in Collections:JABATAN KEJURUTERAAN ELEKTRIK

Files in This Item:
File Description SizeFormat 
Storytelling Style Speech Generation System.pdf7.57 MBAdobe PDFThumbnail
View/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.