the blizzard challenge 2014

14
The Blizzard Challenge 2014 1 Kishore Prahallad, 1 Anandaswarup Vadapalli, 1 Santosh Kesiraju, 2 Hema A. Murthy 3 Swaran Lata, 4 T. Nagarajan, 5 Mahadeva Prasanna, 6 Hemant Patil, 7 Anil Kumar Sao 8 Simon King, 9 Alan W. Black and 10 Keiichi Tokuda 1 Speech and Vision Lab, IIIT Hyderabad, India 2 Department of CSE, IIT Madras, India 3 Department of Electronics and Information Technology, Govt. of India 4 Department of IT, SSN College of Engineering, India 5 Department of EEE, IIT Guwahati, India 6 DAIICT, India 7 School of Computing and Electrical Engineering, IIT Mandi, India 8 Center for Speech Technology Research, University of Edinburgh, UK 9 Language Technologies Institute, Carnegie Mellon University, USA 10 Department of Computer Science, Nagoya Institute of Technology, Japan Abstract The Blizzard challenge 2014 was the tenth annual Blizzard challenge organized by the following group of institutions : IIIT Hyderabad, IIT Madras, DAIICT, SSN College of Engineer- ing, IIT Mandi and IIT Guwahati with support and collabora- tion from DeitY, Government of India. This paper describes the tasks in the Blizzard challenge 2014. The tasks consisted of data from six Indian languages : Assamese, Gujarati, Hindi, Rajasthani, Tamil and Telugu. Seven participants from around the world used the speech data provided as well as the corre- sponding text transcriptions in UTF-8, to build synthetic voices, which were then evaluated by means of listening tests. Index Terms: Blizzard challenge, Speech synthesis, Evaluation of synthetic speech 1. Introduction The Blizzard challenge, originally started by Black and Tokuda [1], is a well established challenge in the field of speech syn- thesis. [1–11] are summary papers which provide information about the previous challenges. These resources can be found on the Blizzard Challenge website 1 . This paper is a summary paper describing the tasks in the Blizzard 2014 challenge. 2. Blizzard 2014 tasks 2.1. Database used Speech and text data for six Indian languages i) Assamese, ii) Gujarati, iii) Hindi, iv) Rajasthani, v) Tamil and vi) Telugu were released. The speech data for each language was 2 hours (sam- pled at 16 KHz), recorded by professional speakers in a high quality studio environment. Along with the speech data the cor- responding text was provoded in UTF-8 format. No other in- formation, like segment labels was provided as part of the chal- lenge. However, there was no restriction on the particpants to learn / use information like phonesets or labels from other re- sources. 1 http://www.festvox.org/blizzard/ For the nature of scripts and sounds of Indian language please refer to [11]. 2.2. Tasks Blizzard challenge 2014 consisted of two tasks, a hub task and and a spoke task. Hub task 2014-IH1 : Participants were asked to build one voice in each language from the provided data, in accordance of the rules of the challenge. The subtasks were numbered from IH1.1 to IH1.6 corresponding to the six languages : IH1.1 (Assamese [AS]), IH1.2 (Gu- jarati [GU]), IH1.3 (Hindi [HI]), IH1.4 (Rajasthani [RJ]), IH1.5 (Tamil [TA]) and IH1.6 (Telugu [TE]). Spoke task 2014-IH2 : Participants had to synthesize multilingual sentences containing Indian language text as well as English. The subtasks were numbered from IH2.1 to IH2.6 corresponding to the six languages : IH2.1 (Assamese [AS]), IH2.2 (Gujarati [GU]), IH2.3 (Hindi [HI]), IH2.4 (Rajasthani [RJ]), IH2.5 (Tamil [TA]) and IH2.6 (Telugu [TE]). For the IH1 task (hub task), the synthetic voices were eval- uated through listening tests on the following test data (for each Indian language) Read speech (RD) - 100 distinct sentences, not a part of the training data Semantically unpredictable sentences (SUS) - 50 distinct sentences not a part of the RD/training data The SUS sentences were prepared in the following man- ner. 50 sentences in each language were randomly selected, and POS tagging was performed on these sentences. The words in each sentence were then reordered as Subject Object Verb Con- juction Subject Object Verb to generate the SUS sentence. For the IH2 task (spoke task), the systems were evaluated through listening tests by synthesizing the following test data (for each Indian language + English combination) Multilingual sentences (ML) - 50 distinct sentences con- taining both Indian language as well as English words.

Upload: danilo-sousa

Post on 15-Aug-2015

131 views

Category:

Education


1 download

TRANSCRIPT

Page 1: The Blizzard Challenge 2014

The Blizzard Challenge 20141Kishore Prahallad, 1Anandaswarup Vadapalli, 1Santosh Kesiraju, 2Hema A. Murthy3Swaran Lata, 4T. Nagarajan, 5Mahadeva Prasanna, 6Hemant Patil, 7Anil Kumar Sao

8Simon King, 9Alan W. Black and 10Keiichi Tokuda

1 Speech and Vision Lab, IIIT Hyderabad, India2 Department of CSE, IIT Madras, India

3 Department of Electronics and Information Technology, Govt. of India4 Department of IT, SSN College of Engineering, India

5 Department of EEE, IIT Guwahati, India6 DAIICT, India

7 School of Computing and Electrical Engineering, IIT Mandi, India8 Center for Speech Technology Research, University of Edinburgh, UK9 Language Technologies Institute, Carnegie Mellon University, USA

10 Department of Computer Science, Nagoya Institute of Technology, Japan

AbstractThe Blizzard challenge 2014 was the tenth annual Blizzardchallenge organized by the following group of institutions : IIITHyderabad, IIT Madras, DAIICT, SSN College of Engineer-ing, IIT Mandi and IIT Guwahati with support and collabora-tion from DeitY, Government of India. This paper describesthe tasks in the Blizzard challenge 2014. The tasks consistedof data from six Indian languages : Assamese, Gujarati, Hindi,Rajasthani, Tamil and Telugu. Seven participants from aroundthe world used the speech data provided as well as the corre-sponding text transcriptions in UTF-8, to build synthetic voices,which were then evaluated by means of listening tests.Index Terms: Blizzard challenge, Speech synthesis, Evaluationof synthetic speech

1. IntroductionThe Blizzard challenge, originally started by Black and Tokuda[1], is a well established challenge in the field of speech syn-thesis. [1–11] are summary papers which provide informationabout the previous challenges. These resources can be foundon the Blizzard Challenge website1. This paper is a summarypaper describing the tasks in the Blizzard 2014 challenge.

2. Blizzard 2014 tasks2.1. Database used

Speech and text data for six Indian languages i) Assamese, ii)Gujarati, iii) Hindi, iv) Rajasthani, v) Tamil and vi) Telugu werereleased. The speech data for each language was 2 hours (sam-pled at 16 KHz), recorded by professional speakers in a highquality studio environment. Along with the speech data the cor-responding text was provoded in UTF-8 format. No other in-formation, like segment labels was provided as part of the chal-lenge. However, there was no restriction on the particpants tolearn / use information like phonesets or labels from other re-sources.

1http://www.festvox.org/blizzard/

For the nature of scripts and sounds of Indian languageplease refer to [11].

2.2. Tasks

Blizzard challenge 2014 consisted of two tasks, a hub task andand a spoke task.

• Hub task 2014-IH1 : Participants were asked to buildone voice in each language from the provided data, inaccordance of the rules of the challenge. The subtaskswere numbered from IH1.1 to IH1.6 corresponding tothe six languages : IH1.1 (Assamese [AS]), IH1.2 (Gu-jarati [GU]), IH1.3 (Hindi [HI]), IH1.4 (Rajasthani [RJ]),IH1.5 (Tamil [TA]) and IH1.6 (Telugu [TE]).

• Spoke task 2014-IH2 : Participants had to synthesizemultilingual sentences containing Indian language textas well as English. The subtasks were numbered fromIH2.1 to IH2.6 corresponding to the six languages :IH2.1 (Assamese [AS]), IH2.2 (Gujarati [GU]), IH2.3(Hindi [HI]), IH2.4 (Rajasthani [RJ]), IH2.5 (Tamil[TA]) and IH2.6 (Telugu [TE]).

For the IH1 task (hub task), the synthetic voices were eval-uated through listening tests on the following test data (for eachIndian language)

• Read speech (RD) - 100 distinct sentences, not a part ofthe training data

• Semantically unpredictable sentences (SUS) - 50 distinctsentences not a part of the RD/training data

The SUS sentences were prepared in the following man-ner. 50 sentences in each language were randomly selected, andPOS tagging was performed on these sentences. The words ineach sentence were then reordered as Subject Object Verb Con-juction Subject Object Verb to generate the SUS sentence.

For the IH2 task (spoke task), the systems were evaluatedthrough listening tests by synthesizing the following test data(for each Indian language + English combination)

• Multilingual sentences (ML) - 50 distinct sentences con-taining both Indian language as well as English words.

Page 2: The Blizzard Challenge 2014

No language tags were provided in the ML sentences. Theparticipants were expected to identify the language from theUnicode code point.

2.3. Participants in the challenge

The participants in the Blizzard challenge 2014 consisted of theseven participants listed in Table 1. To annonimyze the results,the systems are identified using letters, with A denoting naturalspeech, B denoting the baseline system and C to K denoting thesystems submitted by the participants in the challenge. Eachparticipant could submit as many systems as they wished.

Table 1: Participants in Blizzard challenge 2014

Short name Details Synthesis methodNATURAL Natural speechBASE Baseline system HMMNITECH Nagoya Institute of HMM

TechnologyUSTCP National Engineering Hybrid (IH1.3) /

Laboratory of Speech & Language HMM (remaining)Information Processing(Primary system)

CMU Carnegie Mellon University HMMS4A Simple4All project HMM + DNN

consortiumILSP Institute for Language and USS

Speech Processing / InnoeticsIITMS IIT Madras HMM (IH1.3,IH1.4 and IH1.6) /

(Secondary system) USS (remaining)IITMP IIT Madras USS (IH1.3,IH1.4 and IH1.6) /

(Primary system) HMM (remaining)MILE-TTS Dept. of Electrical Engg, USS

Indian Institute of ScienceUSTCS National Engineering HMM

Laboratory of Speech & LanguageInformation Processing(Secondary system)

2.4. Baseline systems

Baseline systems were built for each language using the speakerindependent HTS-2.2 + STRAIGHT scripts2. The data waslabeled at the phone level using the HMM labeling script(EHMM) in FestVox3 [12]. For letter to sound rules a set ofsimple naive first order approximations were used for each lan-guage.

3. EvaluationThe participants were asked to synthesize the complete test set,out of which a subset was used in the listening tests. The lis-tening tests for IH1.1 - IH1.6 consisted of ten sections while thelistening tests for IH2.1 - IH2.6 consisted of five sections. Thedifferent sections of the listening tests are described below.

• Listening tests for IH1.1 - IH1.6

1. two sections for similarity (one section using RDand one section using SUS)

2. seven sections for naturalness (four sections usingRD and three sections using SUS)

3. one section for intelligibility using SUS

• Listening tests for IH2.1 - IH2.6

1. one section for similarity

2http://hts.sp.nitech.ac.jp/?Download3http://www.festvox.org

2. four sections for naturalness

The methodology of scoring in the various sections of thelistening tests are described below.

• Similarity : The listener plays a few samples of the orig-inal speaker and one synthetic sample. The listener thenchooses a response that represented how similar the syn-thetic voice sounded as compared to the original speak-ers voice on a scale from

1 : Sounds like a totally different person

to

5 : Sounds exactly like the same person

• Naturalness : The listener listenes to a sample of syn-thetic speech and chooses a score which represents hownatural or unnatural the sentence sounded on a scale of

1 : Completely Unnatural

to

5 : Completely Natural

• Intelligibility : Listeners listen to an utterance and typein what they hear. Word Error Rate (WER) is computedin the same manner it is computed for speech recognitiontasks.

For the list of changes made in the evaluation portal to en-able the conduct of listening tests in Indian languages, pleaserefer to [11]

4. ResultsThe following listener types were used for the listening tests :

• Paid users

• Online volunteers

Apart from these types of listeners, we also experimentedwith conducting listening tests on Amazon mechanical turk(AMT).

Table 2 shows the statistics of the different listener types forthe tasks.

Table 2: User statistics for the Blizzard 2014 tasks

Task Paid Online AMTUsers volunteers users

IH1.1 + IH1.1 106 09 -IH1.2 + IH2.1 50 0 -IH1.3 + IH2.3 100 09 54IH1.4 + IH2.4 101 09 -IH1.5 + IH2.5 100 09 55IH1.6 + IH2.6 100 06 44

4.1. Results

For the six languages in the IH1 hub task (IH1.1 - IH1.6), Fig-ures 1 to 6 and Figures 7 to 12 show the similarity and natu-ralness results on RD and SUS respectively. The intelligibilityresults for the hub task (IH1.1 - IH1.6) are shown in Figures 13to 18.

For the spoke task (IH2.1 - IH2.6), Figures 19 to 24 showthe similarity and naturalness results on ML.

Page 3: The Blizzard Challenge 2014

For a detailed discussion of the results, please refer to thepapers describing each system submitted by individual partici-pants, available on the Blizzard Challege website.

5. ConclusionsThe conclusions drawn from the results of the Blizzard challege2014 are :

• The high quality audio recordings provided decent per-formances by all systems

• All teams performed better than the baseline system.This can be attributed to the fact that open source toolk-its typically require sufficient tuning to make them workbetter for new/arbitrary languages.

• There does not seem to be much utility in computingWER as a measure of intelligibility for Indian languages.

• Some teams performed better on the ML task as com-pared to RD and SUS.

• Scores obtained from Amazon mechanical turk listenersshow too much noise and variability in the score. Theselisteners can not be used as an alternative to paid listen-ers.

6. References[1] A. W. Black and K. Tokuda, “The Blizzard Challenge - 2005 :

Evaluating corpus-based speech synthesis on common datasets,”in Proceedings of Intespeech 2005, Lisbon, 2005.

[2] C. L. Bennett, “Large scale evaluation of corpus-based synthesiz-ers : Results and lessons from the Blizzard Challenge 2005,” inProceedings of Interspeech 2005, 2005.

[3] C. L. Bennett and A. W. Black, “The Blizzard Challenge 2006,” inBlizzard Challenge Workshop, Interspeech 2006 - ICSLP satelliteevent, 2006.

[4] M. Frazer and S. King, “The Blizzard Challenge 2007,” in Pro-ceedings Blizzard Workshop 2007 (in Proc. SSW6), 2007.

[5] V. Karaiskos, S. King, R. Clark, and C. Mayo, “The BlizzardChallenge 2008,” in Proceedings Blizzard Workshop 2008, 2008.

[6] S. King and V. Karaiskos, “The Blizzard Challenge 2009,” in Pro-ceedings Blizzard Workshop 2009, 2009.

[7] ——, “The Blizzard Challenge 2010,” in Proceedings BlizzardWorkshop 2010, 2010.

[8] ——, “The Blizzard Challenge 2011,” in Proceedings BlizzardWorkshop 2011, 2011.

[9] ——, “The Blizzard Challenge 2012,” in Proceedings BlizzardWorkshop 2012, 2012.

[10] ——, “The Blizzard Challenge 2013,” in Proceedings BlizzardWorkshop 2013, 2013.

[11] K. Prahallad, A. Vadapalli, N. Elluru, G. Mantena, B. Pulugundla,P. Bhaskararao, H. A. Murthy, S. King, V. Karaiskos, and A. W.Black, “The Blizzard Challenge 2013 – Indian Language Tasks,”in Proceedings Blizzard Workshop 2013, 2013.

[12] A. W. Black and K. Lenzo, “Building voices in the festival speechsynthesis system,” 2002, available Online: http://festvox.org/bsv.

[13] R. Clark, M. Podsiadlo, M. Fraser, C. Mayo, and S. King, “Sta-tistical analysis of the Blizzard Challenge 2007 listening test re-sults,” in Proceeding Blizzard Workshop 2007 (in ProceedingsSSW6), 2007.

Page 4: The Blizzard Challenge 2014

●●

●●●

●●

● ●

●●

●●●

●●●

106 106 106 106 106 106 106 106n

A B C D E F G I

12

34

5

RD Mean Opinion Scores (similarity to original speaker) IH1.1 − Paid listeners,

System

Sco

re

●●●

● ●●●●●●●●●●●●●● ●●●●●●●●●●●● ●●●●●●●●●●●●●●●●

424 424 424 424 424 424 424 424n

A B C D E F G I

12

34

5

RD Mean Opinion Scores (naturalness) IH1.1 − Paid listeners

System

Sco

re

Figure 1: Similarity and Naturalness results on RD for IH1.1 (Assamese)

●●●●

●●●●

50 50 50 50 50 50 50 50 50n

A B C D E F G H I

12

34

5

RD Mean Opinion Scores (similarity to original speaker) IH1.2 − Paid listeners,

System

Sco

re

●●●

●●●●●●●

●●●●●● ●●●●●●●● ●●●●●●●●●●●

●●●●●●●●●

200 200 200 200 200 200 200 200 200n

A B C D E F G H I

12

34

5

RD Mean Opinion Scores (naturalness) IH1.2 − Paid listeners

System

Sco

re

Figure 2: Similarity and Naturalness results on RD for IH1.2 (Gujarati)

Page 5: The Blizzard Challenge 2014

●●

●●●●●●

● ●●●●●●● ●●●●●

100 100 100 100 100 100 100 100 100 100n

A B C D E F G H I K

12

34

5

RD Mean Opinion Scores (similarity to original speaker) IH1.3 − Paid listeners,

System

Sco

re

●●●●

●●●●●

●●●

●●●●●●●●●●●●●●●●●●●●

●●●●●●●●●●●●● ●●●●●● ●●●●●●●●●●●●●●

400 400 400 400 400 400 400 400 400 400n

A B C D E F G H I K

12

34

5

RD Mean Opinion Scores (naturalness) IH1.3 − Paid listeners

System

Sco

re

Figure 3: Similarity and Naturalness results on RD for IH1.3 (Hindi)

●●

●●●

●●●●●●●

●●●● ●●●●●●●●

●●

●●

●●●●●●●●

101 101 101 101 101 101 101 101 101n

A B C D E F G H I

12

34

5

RD Mean Opinion Scores (similarity to original speaker) IH1.4 − Paid listeners,

System

Sco

re

●●

●●●

●●●●

●●

●●●●●●●●●●

●●

●●●●●●

●●●●

●●●●●●●●●●●●●●●●

●●●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●●●●●●●●●

404 404 404 404 404 404 404 404 404n

A B C D E F G H I

12

34

5

RD Mean Opinion Scores (naturalness) IH1.4 − Paid listeners

System

Sco

re

Figure 4: Similarity and Naturalness results on RD for IH1.4 (Rajasthani)

Page 6: The Blizzard Challenge 2014

●●●●

●●

100 100 100 100 100 100 100 100 100 100n

A B C D E F G H I J

12

34

5

RD Mean Opinion Scores (similarity to original speaker) IH1.5 − Paid listeners,

System

Sco

re

●●●

●●

●●●●●●

●●

●●●

●●

●●

●●

●●●●●

●●●

●●

●●●●●●●

●●

●●●●●●

●●

●●●

●●

●●

●●●●●

●●●●●●●●●●●●●●●●●●●●

●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●●●●●●●●

400 400 400 400 400 400 400 400 400 400n

A B C D E F G H I J

12

34

5

RD Mean Opinion Scores (naturalness) IH1.5 − Paid listeners

System

Sco

re

Figure 5: Similarity and Naturalness results on RD for IH1.5 (Tamil)

●●

●●●

100 100 100 100 100 100 100 100 100n

A B C D E F G H I

12

34

5

RD Mean Opinion Scores (similarity to original speaker) IH1.6 − Paid listeners,

System

Sco

re

●●●●●●●●●●●●●●●

●●●●●

●●●●●●●

●●●●●● ●●●●●●●●●●●●●

●●●●●●●

●●●

●●

●●

●●●●●●●●●●●

●●●●●●

400 400 400 400 400 400 400 400 400n

A B C D E F G H I

12

34

5

RD Mean Opinion Scores (naturalness) IH1.6 − Paid listeners

System

Sco

re

Figure 6: Similarity and Naturalness results on RD for IH1.6 (Telugu)

Page 7: The Blizzard Challenge 2014

●●

●●●

●●●

106 106 106 106 106 106 106 106n

A B C D E F G I

12

34

5

SUS Mean Opinion Scores (similarity to original speaker) IH1.1 − Paid listeners

System

Sco

re

●●

●●●●

●●

●●●●●●●●●●●●

●●●

●●●●●●●●●●●●●●●

●●●●

●●

●●● ●●

●●●●●●

●●●●●●●

●●●●●●●

●●●●●● ●●●●●●●●●●● ●●

●●●●●●●

●●●●●●●●●●●

318 318 318 318 318 318 318 318n

A B C D E F G I

12

34

5

SUS Mean Opinion Scores (naturalness) IH1.1 − Paid listeners

System

Sco

re

Figure 7: Similarity and Naturalness results on SUS for IH1.1 (Assamese)

●●● ●

●●●● ●●

50 50 50 50 50 50 50 50 50n

A B C D E F G H I

12

34

5

SUS Mean Opinion Scores (similarity to original speaker) IH1.2 − Paid listeners

System

Sco

re

●●

●●●

●● ●●●●●●

●●●●

●●●

150 150 150 150 150 150 150 150 150n

A B C D E F G H I

12

34

5

SUS Mean Opinion Scores (naturalness) IH1.2 − Paid listeners

System

Sco

re

Figure 8: Similarity and Naturalness results on SUS for IH1.2 (Gujarati)

Page 8: The Blizzard Challenge 2014

●●●●●● ●●●●

100 100 100 100 100 100 100 100 100 100n

A B C D E F G H I K

12

34

5

SUS Mean Opinion Scores (similarity to original speaker) IH1.3 − Paid listeners

System

Sco

re

●●●●

●●

●●

●●

●●●

●●●●●● ●●●●●●●●●●●

●●●●●●●●●●●●●●●●● ●●●●●●●●●

300 300 300 300 300 300 300 300 300 300n

A B C D E F G H I K

12

34

5

SUS Mean Opinion Scores (naturalness) IH1.3 − Paid listeners

System

Sco

re

Figure 9: Similarity and Naturalness results on SUS for IH1.3 (Hindi)

●●●

101 101 101 101 101 101 101 101 101n

A B C D E F G H I

12

34

5

SUS Mean Opinion Scores (similarity to original speaker) IH1.4 − Paid listeners

System

Sco

re

●●

●●

●●●●●●●●●●●●●●●●●●●●●●●

●●●●●●●●●●●●●●●●●●●●

●●●●●●●●●●●●●●●

●●●

●●●●●●●●●●●●

●●●●●●●●●●●●

●●●●●● ●●●●●●●●●●●●●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●

303 303 303 303 303 303 303 303 303n

A B C D E F G H I

12

34

5

SUS Mean Opinion Scores (naturalness) IH1.4 − Paid listeners

System

Sco

re

Figure 10: Similarity and Naturalness results on SUS for IH1.4 (Rajasthani)

Page 9: The Blizzard Challenge 2014

100 100 100 100 100 100 100 100 100 100n

A B C D E F G H I J

12

34

5

SUS Mean Opinion Scores (similarity to original speaker) IH1.5 − Paid listeners

System

Sco

re

●●●●●●●

●●

●●●

●●

●●●●●●●●●●●●

●●●●●●●●●●●●

●●●●●●●●●●●●

300 300 300 300 300 300 300 300 300 300n

A B C D E F G H I J

12

34

5

SUS Mean Opinion Scores (naturalness) IH1.5 − Paid listeners

System

Sco

re

Figure 11: Similarity and Naturalness results on SUS for IH1.5 (Tamil)

●●

●● ●●

●●●●●●● ●●●●●●●●

100 100 100 100 100 100 100 100 100n

A B C D E F G H I

12

34

5

SUS Mean Opinion Scores (similarity to original speaker) IH1.6 − Paid listeners

System

Sco

re

●●●●

●●●

●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●

●●●

●●●●

●●●

●●●●●●●● ●●●●●●●●●

●●●●●●●●●●● ●●●●●●● ●●●●

●●●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●

●●●● ●●●●●●●●●●●

●●●

300 300 300 300 300 300 300 300 300n

A B C D E F G H I

12

34

5

SUS Mean Opinion Scores (naturalness) IH1.6 − Paid listeners

System

Sco

re

Figure 12: Similarity and Naturalness results on SUS for IH1.6 (Telugu)

Page 10: The Blizzard Challenge 2014

A B C D E F G I

05

1015

2025

3035

4045

5055

6065

7075

8085

9095

100

106 101 102 99 103 102 103 104n

SUS Word error rate (IH1.1 Paid listeners)

System

WE

R (

%)

Figure 13: Intelligibility results on SUS for IH1.1 (As-samese)

A B C D E F G H I

05

1015

2025

3035

4045

5055

6065

7075

8085

9095

100

49 49 48 47 48 50 48 48 49n

SUS Word error rate (IH1.2 Paid listeners)

System

WE

R (

%)

Figure 14: Intelligibility results on SUS for IH1.2 (Gujarati)

A B C D E F G H I K

05

1015

2025

3035

4045

5055

6065

7075

8085

9095

100

99 99 95 100 100 100 99 100 99 100n

SUS Word error rate (IH1.3 Paid listeners)

System

WE

R (

%)

Figure 15: Intelligibility results on SUS for IH1.3 (Hindi)

A B C D E F G H I

05

1015

2025

3035

4045

5055

6065

7075

8085

9095

100

101 46 99 100 101 101 101 100 101n

SUS Word error rate (IH1.4 Paid listeners)

System

WE

R (

%)

Figure 16: Intelligibility results on SUS for IH1.4 (Ra-jasthani)

Page 11: The Blizzard Challenge 2014

A B C D E F G H I J

05

1015

2025

3035

4045

5055

6065

7075

8085

9095

100

96 95 85 98 99 98 99 92 97 95n

SUS Word error rate (IH1.5 Paid listeners)

System

WE

R (

%)

Figure 17: Intelligibility results on SUS for IH1.5 (Tamil)

A B C D E F G H I

05

1015

2025

3035

4045

5055

6065

7075

8085

9095

100

100 89 86 98 98 100 99 100 95n

SUS Word error rate (IH1.6 Paid listeners)

System

WE

R (

%)

Figure 18: Intelligibility results on SUS for IH1.6 (Telugu)

Page 12: The Blizzard Challenge 2014

●●

●●

●●

106 106 106 106 106n

A B C D E

12

34

5

ML Mean Opinion Scores (similarity to original speaker) IH2.1 − Paid listeners

System

Sco

re

●●●●●●●●●

●●

●●●●●●●●●●●●●●●●●●●● ●●●●●●●●●

●●●●●●●●

●●●●●

●●●

●●●

●●●●●●

●●●●●●●● ●●●●●●

424 424 424 424 424n

A B C D E

12

34

5

ML Mean Opinion Scores (naturalness) IH2.1 − Paid listeners

System

Sco

re

Figure 19: Similarity and Naturalness results on ML for IH2.1 (Assamese)

●●● ●●

50 50 50 50 50n

A B C D E

12

34

5

ML Mean Opinion Scores (similarity to original speaker) IH2.2 − Paid listeners

System

Sco

re

●●●●●●●●●●● ●●●●

200 200 200 200 200n

A B C D E

12

34

5

ML Mean Opinion Scores (naturalness) IH2.2 − Paid listeners

System

Sco

re

Figure 20: Similarity and Naturalness results on ML for IH2.2 (Gujarati)

Page 13: The Blizzard Challenge 2014

●●●●●●

●●●●●●●●●●●●

●●●

100 100 100 100 100 100 100n

A B C D E F K

12

34

5

ML Mean Opinion Scores (similarity to original speaker) IH2.3 − Paid listeners

System

Sco

re

●●

●●

●●●●

●●●

●●

●●●

●●●●●

●●

●●●

●●●●●●●

●●●●● ●●●●●●●●●●●●

300 300 300 300 300 300 300n

A B C D E F K

12

34

5

ML Mean Opinion Scores (naturalness) IH2.3 − Paid listeners

System

Sco

re

Figure 21: Similarity and Naturalness results on ML for IH2.3 (Hindi)

●●●●●●●●

● ●●●● ●●●●●●● ●●●●●●

101 101 101 101 101 101n

A B C D E F

12

34

5

ML Mean Opinion Scores (similarity to original speaker) IH2.4 − Paid listeners

System

Sco

re

●●

●●●

●●

●●●●

●●

●●●

●●●

●●●●

●●●

●●

●●

●●●

●●●●●●●●●●●●●●●

●●●●●●●●●●●●●● ●●●●●●●●●●● ●●●●●●●●●● ●●●●●●●●●●●●●●●●●●●●●●●●

404 404 404 404 404 404n

A B C D E F

12

34

5

ML Mean Opinion Scores (naturalness) IH2.4 − Paid listeners

System

Sco

re

Figure 22: Similarity and Naturalness results on ML for IH2.4 (Rajasthani)

Page 14: The Blizzard Challenge 2014

100 100 100 100 100 100n

A B C D E J

12

34

5

ML Mean Opinion Scores (similarity to original speaker) IH2.5 − Paid listeners

System

Sco

re

●●

●●●

●●

●●●●●

●●

●●●●●●●●●●●●●●●●●●●●● ●●●●●●●●●●●●●●●●●●●●●●●●

400 400 400 400 400 400n

A B C D E J

12

34

5

ML Mean Opinion Scores (naturalness) IH2.5 − Paid listeners

System

Sco

re

Figure 23: Similarity and Naturalness results on ML for IH2.5 (Tamil)

●●●

100 100 100 100 100 100n

A B C D E F

12

34

5

ML Mean Opinion Scores (similarity to original speaker) IH2.6 − Paid listeners

System

Sco

re

●●●

●●●●●●●●●●●

●●●●●●●●●●●●

●●●●●●● ●●●●●●●●●●●●

●●●●●●●●●●

400 400 400 400 400 400n

A B C D E F

12

34

5

ML Mean Opinion Scores (naturalness) IH2.6 − Paid listeners

System

Sco

re

Figure 24: Similarity and Naturalness results on ML for IH2.6 (Telugu)