TY - GEN
T1 - Comparison of native and nonnative speakers' perspective in animated text visualization tool
AU - Binti Samsudin, Nur Syafikah
AU - Mano, Kazunori
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2016/1/5
Y1 - 2016/1/5
N2 - Media conversion technologies such as speech recognition and speech synthesis have gained a lot of attention in recent years. They are applied in various human communication tools through smart phones and personal computers such as in language learning system. However, the learner still have loss in naturalness pronunciation problem because of effects on mother tongue interference. Although the learner can realize that his/her speech is different from the trainer's, the learner still cannot detect or check the precise wrong part in their utterances speech. We indicate these differences by visualizing a learner's wrong and correct pronunciation with speech-to-animated text visualization tool. In this study, we focused on the media conversion process between speech prosodic information and animated text information by using AHP method as the mapping method. The pairwise comparison between both speech and text information were conducted and evaluated by Japanese native and nonnative speakers. Here, we investigated the comparison between native and nonnative speakers' perspective and determined the ideal matching elements between attributes of speech and text information.
AB - Media conversion technologies such as speech recognition and speech synthesis have gained a lot of attention in recent years. They are applied in various human communication tools through smart phones and personal computers such as in language learning system. However, the learner still have loss in naturalness pronunciation problem because of effects on mother tongue interference. Although the learner can realize that his/her speech is different from the trainer's, the learner still cannot detect or check the precise wrong part in their utterances speech. We indicate these differences by visualizing a learner's wrong and correct pronunciation with speech-to-animated text visualization tool. In this study, we focused on the media conversion process between speech prosodic information and animated text information by using AHP method as the mapping method. The pairwise comparison between both speech and text information were conducted and evaluated by Japanese native and nonnative speakers. Here, we investigated the comparison between native and nonnative speakers' perspective and determined the ideal matching elements between attributes of speech and text information.
KW - Animated text
KW - Nonlinguistic information
KW - Paralinguistic information
KW - Speech visualization
UR - http://www.scopus.com/inward/record.url?scp=84962199661&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84962199661&partnerID=8YFLogxK
U2 - 10.1109/TENCON.2015.7372934
DO - 10.1109/TENCON.2015.7372934
M3 - Conference contribution
AN - SCOPUS:84962199661
T3 - IEEE Region 10 Annual International Conference, Proceedings/TENCON
BT - TENCON 2015 - 2015 IEEE Region 10 Conference
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 35th IEEE Region 10 Conference, TENCON 2015
Y2 - 1 November 2015 through 4 November 2015
ER -