Introduction   Many 3-D pronunciation tutors with both internal and external articulator movements have been implemented and applied to computer-aided.

Slides:



Advertisements
Similar presentations
Project VIABLE: Behavioral Specificity and Wording Impact on DBR Accuracy Teresa J. LeBel 1, Amy M. Briesch 1, Stephen P. Kilgus 1, T. Chris Riley-Tillman.
Advertisements

Eye Movements of Younger and Older Drivers Professor: Liu Student: Ruby.
Background POP displays are critical because 68% of all buying decisions are made with no prior planning (Stahlberg & Maila, 2010). Effective POP displays.
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
The Use of Eye Tracking Technology in the Evaluation of e-Learning: A Feasibility Study Dr Peter Eachus University of Salford.
Comparison of two eye tracking devices used on printed images Barbora Komínková The Norwegian Color Research Laboratory Faculty of Computer Science and.
ICT TEACHERS` COMPETENCIES FOR THE KNOWLEDGE SOCIETY
Chinese People's Self-Construal and Their Social Participation in Beijing Olympic Games(BOG) Bing Wang Institute of Sociology Chinese Academy of Social.
Multimodality and Activity Theory Methodological issues in their combination Dr. Mohammed Alhuthali Taif University, Saudi Arabia
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Evaluating Eye Movement Differences when Processing Subtitles Andrew T. Duchowski COMPUTER SCIENCE CLEMSON UNIVERSITY Abstract Our experimental.
1 Integrating Google Apps for Education to Business English Student Trainees’ On-the-Job Training English Reports Asst.Prof. Phunsuk Kannarik.
AN ITERATIVE METHOD FOR MODEL PARAMETER IDENTIFICATION 4. DIFFERENTIAL EQUATION MODELS E.Dimitrova, Chr. Boyadjiev E.Dimitrova, Chr. Boyadjiev BULGARIAN.
Year Review Nancy Rader May 13, esearch Emotion and Working Memory Temperament Infant Perception Attention and Early Language.
Authors: Xu Cheng, Haitao Li, Jiangchuan Liu School of Computing Science, Simon Fraser University, British Columbia, Canada. Speaker : 童耀民 MA1G0222.
A MULTIMEDIA APPLICATION FOR THE TEACHING OF THE MODULE "WATER" Anna Thysiadou 1, Sofoklis Christoforidis 2, Panagiotis Jannakoudakis 1 1 Aristotle University.
Acoustic Properties of Taiwanese High School Students ’ Stress in English Intonation Advisor: Dr. Raung-Fu Chung Student: Hong-Yao Chen.
Integrating Handheld Computers in a Middle-School Science Curriculum Jason M. Williams Georgia Institute of Technology.
Eye Tracking In Evaluating The Effectiveness OF Ads Guide : Dr. Andrew T. Duchowski.
Semantic v.s. Positions: Utilizing Balanced Proximity in Language Model Smoothing for Information Retrieval Rui Yan†, ♮, Han Jiang†, ♮, Mirella Lapata‡,
An Analysis of Advertisement Perception through Eye Tracking William A. Hill Physics Youngstown State University Zach C. Joyce Computer.
An Eyetracking Analysis of the Effect of Prior Comparison on Analogical Mapping Catherine A. Clement, Eastern Kentucky University Carrie Harris, Tara Weatherholt,
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Parking Study Analysis Prepared by : Mashel ALDaajam Mohammed ALMasand and Mohammed ALMannaa.
Distinguishing humans from robots in web search logs preliminary results using query rates and intervals Omer Duskin Dror G. Feitelson School of Computer.
Chinese Learners’ Perception and Production of the vowels: /e/, /ei/, /o/, & /ou/ in English by Contrastive Analysis 研究生 : 張悅寧 報告人 : NA2C0003 傅學琳 WHO WHAT.
Using Blog Properties to Improve Retrieval Gilad Mishne (ICWSM 2007)
A CONTENT ANALYTIC COMPARISON OF LEARNING PROCESSES IN ONLINE and FACE- TO-FACE CASE STUDY DISCUSSIONS by Heckman and Annabi (2005)
Introduction   Many 3-D pronunciation tutors with both internal and external articulator movements have been implemented and applied to computer-aided.
Authors: Jiang Xie, Ian F. Akyildiz
The influence of visual strategies used by National Water Polo goalies on decision time and accuracy improvement through a video based perceptual training.
Visualization of Eye Gaze Data using Heat Maps
Project No EPP EL-EPPKA2-CBHE-JP ( ) UES TEAM
Assessing Students' Understanding of the Scientific Process Amy Marion, Department of Biology, New Mexico State University Abstract The primary goal of.
Lab Roles and Lab Report
Databases and Database Management Systems Chapter 9
Do they learn differently
Parts of an Academic Paper
GLoCALL & PCBET 2017 Joint Conference, 7-9 September 2017 at Universiti Teknologi Brunei, Brunei Darussalam, Presented at Room 1, 11:00-11:30. Effect of.
Competition, financial innovation and commercial
MICHAEL NEEL, University of Houston
Duration of presentat ion
Angry Faces Capture Attention But Do They Hold It?
Authoring Directed Gaze for Full-Body Motion Capture
An Exploration of the Perceived Femininity of Language Learning
Single-Neuron Correlates of Atypical Face Processing in Autism
Neuronal Correlates of Metacognition in Primate Frontal Cortex
Course Access Frequencies and Student Grades
Sign critical appraisal course: exercise 2
Chenguang Zheng, Kevin Wood Bieri, Yi-Tse Hsiao, Laura Lee Colgin 
Volume 27, Issue 2, Pages (January 2017)
Colin J. Palmer, Colin W.G. Clifford  Current Biology 
Feature- and Order-Based Timing Representations in the Frontal Cortex
Attentional Modulations Related to Spatial Gating but Not to Allocation of Limited Resources in Primate V1  Yuzhi Chen, Eyal Seidemann  Neuron  Volume.
Cultural Confusions Show that Facial Expressions Are Not Universal
Huihui Zhou, Robert Desimone  Neuron 
Neural Mechanisms of Visual Motion Perception in Primates
Volume 56, Issue 1, Pages (October 2007)
Sharon C. Furtak, Omar J. Ahmed, Rebecca D. Burwell  Neuron 
Xiaomo Chen, Marc Zirnsak, Tirin Moore  Cell Reports 
Contextualised MALL Comparison of Chinese Learning Activities Based on Country (Target/Non-Target) and Language Learning Orientation (Generic/Dedicated)
Chapter 10 Introduction to the Analysis of Variance
Encoding of Stimulus Probability in Macaque Inferior Temporal Cortex
Retrieval Performance Evaluation - Measures
Reading and Evaluating Research Reports
Repetition suppression: Interaction between trials without (prime) versus with (target) context (C) and repeated versus non-repeated context (i.e., repetition.
EAST MDSplus Log Data Management System
Manuel Jan Roth, Matthis Synofzik, Axel Lindner  Current Biology 
Embry-Riddle Aeronautical University, College of Aviation
Yingze Wang and Shi-Kuo Chang University of Pittsburgh
Presentation transcript:

EVALUATION OF A MULTIMODAL 3-D PRONUNCIATION TUTOR FOR LEARNING MANDARIN AS A SECOND LANGUAGE: AN EYE-TRACKING STUDY Ying Zhou1,2, Fei Chen2, Hui Chen3, Lan Wang2, Nan Yan2 1 School of Information Engineering, Wuhan University of Technology, Wuhan, China 2 Key Laboratory of Human-Machine Intelligence-Synergy Systems, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China 3 Institute of Software, Chinese Academy of Sciences, Beijing, China

Introduction   Many 3-D pronunciation tutors with both internal and external articulator movements have been implemented and applied to computer-aided language learning (CALL).  In our previous study, a multimodal 3-D articulation system has been presented that incorporated an airflow model into the 3-D articulatory model to produce the airflow in accordance with articulator movements of Mandarin pronunciation. Data of participants’ eye-movement behavior showed a correlation between the duration of eye fixations and the depth of learning.  In the present study, by using eye-tracking methodology, evaluation of the multimodal 3-D Mandarin pronunciation tutor was conducted in comparison with real human teacher to penetrate whether the 3-D talking head can work well during the process of language learning. 

 Through analyzing eye-tracking measures, three research questions (RQ) were put forward and investigated: RQ1: Comparatively, which is more attractive to learners, HF or 3-D? RQ2: Under different presentation conditions, do learners pay different attentions to the areas of interest (AOI, see Data Analyses for more details)? Which is directly related to language learning? RQ3: During the process of learning, will learners focus on airflow information which is useful to increase the syllable identification accuracy by illustrating the airflow differences between the confusable consonants?

Participants, materials and procedure  Methods  Participants, materials and procedure   Fourteen non-Chinese speakers (10 males) between 20 and 31 years of age.  Eye movement data of the participants was collected non-intrusively by RED 5 Eye Tracker (SMI Technology, Germany), which was integrated within the panels of the monitor.  The 16 Mandarin syllables were presented under two conditions of HF and 3-D, totaling 32 videos, each with a front view first and then with a corresponding profile view. 

1) ‘entry time’ (ET), which means the duration from start of the trial to the first hit of one special area. The shorter the ET, the more interest for the area was shown. 2) ‘fixation count’ (FC), which is defined as the total number of fixation lasting more than 100 ms inside the area of interest (AOI), reflecting the absolute attention during the process of learning. 3) ‘proportion of fixation duration’ (POFD), indicating the ratio of fixation duration inside the AOI to the duration of whole video screen, reflecting the relative attention during learning. ► The chosen area of interest (AOI) with front and profile view corresponds to the area inside the dashed red rectangle.

Comparison of entry time (ET) into HF and 3-D videos Results  Comparison of entry time (ET) into HF and 3-D videos The similar results went to the the average ET into different presentation conditions, which could be observed more visually from figure as follows.. ET into 3-D condition was significantly shorter than that into HF condition with the front view [F (1, 26) = 4.687; p < 0.05], and with the profile view [F (1, 26) = 24.147; p < 0.001].  

Eye-tracking measures of AOI  Eye-tracking measures of AOI Post hoc analysis indicated that with the front view, presentation condition showed no significant effect (p = 0.229). With the profile view, however, ET was much shorter in 3-D (377.58 ms) than in HF (487.80 ms) condition (p < 0.01) 

 With the front view, presentation condition showed no significant effect (p = 0.073), while with the profile view, the FC of AOI in 3-D (5.41) was higher (p < 0.001) compared with that in HF condition (4.35) (see Table 1 and Fig. 4)

 Post hoc analysis indicated that the POFD of AOI in 3-D was significantly higher than that in HF condition both with front view and profile view (all ps < 0.01)

Distribution of attention inside the AOI of profile 3-D video  Distribution of attention inside the AOI of profile 3-D video The average FC between the two subareas (see Figure) was not different from one another (t = 1.916, p = 0.066). 

Discussion  Results indeed showed that ET into 3-D videos was much shorter than the HF videos with both the front and profile view, indicating that non-Chinese learners showed more interests in our 3-D Mandarin pronunciation tutor.  The results showed that learners put more relative attention (i.e., higher POFD) to the lip movement in the 3-D tutors. Moreover, the results of three eye-tracking measures of AOI (i.e., ET, FC, and POFD) all showed that 3-D pronunciation tutor with a transparent profile view triumphed over the real human teacher by effectively delivering articulator movement and airflow information (see Heat map ).  The results of our current study proved that, while watching the 3-D pronunciation tutor, the allocation of absolute attention was equably distributed among the entire AOI with the transparent profile view, rather than concentrating merely on internal articulators

Conclusions  With a shorter entry time into the 3-D videos, non-Chinese learners showed more preference for our 3-D Mandarin pronunciation tutor.  Learners observed the lip movement of 3-D tutor for a relatively longer time with a front view.  With the transparent profile view, our multimodal 3-D pronunciation tutor exhibited greater advantage of delivering articulator movements and airflow information.  Learning from the 3-D tutor in a transparent profile view, learners kept focusing on the entire AOI (including the internal articulators, mouth and airflow) which is important for Mandarin articulation learning.

The end Thank you!