1 Evaluating the Effect of Predicting Oral Reading Miscues Satanjeev Banerjee, Joseph Beck, Jack Mostow Project LISTEN (www.cs.cmu.edu/~listen) Carnegie.

Slides:



Advertisements
Similar presentations
Estimating the detector coverage in a negative selection algorithm Zhou Ji St. Jude Childrens Research Hospital Dipankar Dasgupta The University of Memphis.
Advertisements

Spelling Correction for Search Engine Queries Bruno Martins, Mario J. Silva In Proceedings of EsTAL-04, España for Natural Language Processing Presenter:
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Testing: Principles and Techniques  Tests are “inappropriate, mysterious, unreal, subjective, and unstructured.”  Certain basic questions need to be.
Speech Recognition Part 3 Back end processing. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
USING LINKING FEATURES IN LEARNING NON-PARAMETRIC PART MODELS * Ammar Kamal Hattab ENGN2560 Final Project Presentation May 17, 2013 * Leonid Karlinsky,
Carnegie Mellon 1. Toward Exploiting EEG Input in a Reading Tutor Jack Mostow, Kai-min Chang, and Jessica Nelson Project LISTEN
Detecting Prosody Improvement in Oral Rereading Minh Duong and Jack Mostow Project LISTEN Carnegie Mellon University The research.
On-line Learning with Passive-Aggressive Algorithms Joseph Keshet The Hebrew University Learning Seminar,2004.
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
The Relevance Model  A distribution over terms, given information need I, (Lavrenko and Croft 2001). For term r, P(I) can be dropped w/o affecting the.
Finding Advertising Keywords on Web Pages Scott Wen-tau YihJoshua Goodman Microsoft Research Vitor R. Carvalho Carnegie Mellon University.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Aims of session Making reading fun Early reading Developing reading
® Automatic Scoring of Children's Read-Aloud Text Passages and Word Lists Klaus Zechner, John Sabatini and Lei Chen Educational Testing Service.
How Spread Works. Spread Spread stands for Speech and Phoneme Recognition as Educational Aid for the Deaf and Hearing Impaired Children It is a game used.
Improving the Help Selection Policy in a Reading Tutor that Listens Cecily Heiner, Joseph E. Beck, Jack Mostow Project LISTEN
Jeopardy Proofreading 1 Proofreading 2 Proofreading Marks Proofreading Marks 4 Proofreading
Recognition of spoken and spelled proper names Reporter : CHEN, TZAN HWEI Author :Michael Meyer, Hermann Hild.
Experimentation Duration is the most significant feature with around 40% correlation. Experimentation Duration is the most significant feature with around.
Presented by: Fang-Hui Chu Boosting HMM acoustic models in large vocabulary speech recognition Carsten Meyer, Hauke Schramm Philips Research Laboratories,
ELIS-DSSP Sint-Pietersnieuwstraat 41 B-9000 Gent SPACE symposium - 6/2/091 Language modelling (word FST) Operational model for categorizing mispronunciations.
Experimentation Duration is the most significant feature with around 40% correlation. Experimentation Duration is the most significant feature with around.
Assessment of Phonology
Detection, Classification and Tracking in a Distributed Wireless Sensor Network Presenter: Hui Cao.
Boosting Training Scheme for Acoustic Modeling Rong Zhang and Alexander I. Rudnicky Language Technologies Institute, School of Computer Science Carnegie.
Developing an automated assessment tool for children’s oral reading Leen Cleuren March
What have we learned?. What is a database? An organized collection of related data.
Comparative Study of Name Disambiguation Problem using a Scalable Blocking-based Framework Byung-Won On, Dongwon Lee, Jaewoo Kang, Prasenjit Mitra JCDL.
Evaluating Results of Learning Blaž Zupan
Voice Activity Detection based on OptimallyWeighted Combination of Multiple Features Yusuke Kida and Tatsuya Kawahara School of Informatics, Kyoto University,
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Learning to Detect Faces A Large-Scale Application of Machine Learning (This material is not in the text: for further information see the paper by P.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
13 th October Vocabulary We use the correct terminology with the children right from Early Years. It may sound complicated but it actually makes.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Experimentation Duration is the most significant feature with around 40% correlation. Experimentation Duration is the most significant feature with around.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
On using context for automatic correction of non-word misspellings in student essays Michael Flor Yoko Futagi Educational Testing Service 2012 ACL.
Kindergarten Language Arts
Brian Lukoff Stanford University October 13, 2006.
Using Conversational Word Bursts in Spoken Term Detection Justin Chiu Language Technologies Institute Presented at University of Cambridge September 6.
Combining Evolutionary Information Extracted From Frequency Profiles With Sequence-based Kernels For Protein Remote Homology Detection Name: ZhuFangzhi.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Phone-Level Pronunciation Scoring and Assessment for Interactive Language Learning Speech Communication, 2000 Authors: S. M. Witt, S. J. Young Presenter:
S1S1 S2S2 S3S3 8 October 2002 DARTS ATraNoS Automatic Transcription and Normalisation of Speech Jacques Duchateau, Patrick Wambacq, Johan Depoortere,
ASSESSING SEARCH TERM STRENGTH IN SPOKEN TERM DETECTION Amir Harati and Joseph Picone Institute for Signal and Information Processing, Temple University.
Author :K. Thambiratnam and S. Sridharan DYNAMIC MATCH PHONE-LATTICE SEARCHES FOR VERY FAST AND ACCURATE UNRESTRICTED VOCABULARY KEYWORD SPOTTING Reporter.
Writing progression from Reception to Year 1. End of year expectations. Expected: Children use their phonic knowledge to write words in ways which match.
ASSESSING SEARCH TERM STRENGTH IN SPOKEN TERM DETECTION Amir Harati and Joseph Picone Institute for Signal and Information Processing, Temple University.
A Text-free Approach to Assessing Nonnative Intonation Joseph Tepperman, Abe Kazemzadeh, and Shrikanth Narayanan Signal Analysis and Interpretation Laboratory,
Using Speech Recognition to Predict VoIP Quality
Wei Chen, Jack Mostow Gregory Aist Project LISTEN
Speaker : chia hua Authors : Long Qin, Ming Sun, Alexander Rudnicky
Challenges in Creating an Automated Protein Structure Metaserver
Evaluating Results of Learning
Erasmus University Rotterdam
Leveled Reading Assessment Training
Kindergarten Scope & Sequence Unit 10: School’s Out!
Do-Gil Lee1*, Ilhwan Kim1 and Seok Kee Lee2
Micro-analysis of Fluency Gains in a Reading Tutor that Listens:
Detecting Prosody Improvement in Oral Rereading
Text-To-Speech Tool For Research Paper
Automating Early Assessment of Academic Standards for Very Young Native and Non-Native Speakers of American English better known as The TBALL Project.
Independent versus Computer-Guided Oral Reading:
An Embedded Experiment to Evaluate the Effectiveness of Vocabulary Previews in an Automated Reading Tutor Jack Mostow, Joe Beck, Juliet Bey, Andrew Cuneo,
Jack Mostow* and Joseph Beck Project LISTEN (
Educational Data Mining Success Stories
Experimenter-defined measures in a Reading Tutor that Listens
Presentation transcript:

1 Evaluating the Effect of Predicting Oral Reading Miscues Satanjeev Banerjee, Joseph Beck, Jack Mostow Project LISTEN ( Carnegie Mellon University Funding: NSF IERI

2 Why Predict Miscues? Reading Tutor helps children learn to read. Speech recognizer listens for miscues (reading errors) –E.g.: listen for “hat” if sentence to be read has word “hate” Accurate miscue prediction helps miscue detection.

3 Real Word Substitutions Miscues = substitutions, omissions, insertions Real word substitution = misread target word as another word –E.g. read “hat” instead of “hate” Most miscues are real word substitutions ICSLP-02: predicted real word substitutions Here: evaluate effect on substitution detection

4 How Evaluate Substitution Detection? What child saidIhathavingmilkeverya Correct textIhatehavingmilkeveryday What ASR heard Ihatehavingmilleveryhate substitution undetectedfalse alarmsubstitution detected substitution 1 2 Substitution detection rate = # substitutions detected # substitutions child made = 1 4 False alarm rate = # false alarms # words correctly read =

5 Evaluation Data Sentences read by 25 children aged 6 to 10 Correctly readIncorrectly read Content tokens 5, Function tokens 3,166147

6 Rote Method Uses the University of Colorado miscue database. For each target word –Sort substitutions by # children who made them. –Predict that the top n substitutions will reoccur, for this word. Model type Substitution detection rate False alarm rate No predicted substitutions %2.42 % Top %2.88 % Top %3.27 %

7 Extrapolative Method Predict the probability that a word is a likely substitution for another word –Pr ( substitution “hat” | target “hate”) Use machine learning to induce a classifier Train using University of Colorado miscue database. Some features (more in paper) Candidate substitution Target word Spelling edit distance = 1H.A.T.EH.A.T Phonetic distance = 1/HH EY T//HH AE T/ Rank in descending sorted frequency table

8 Extrapolative Method cont’d Given a target word, predict substitution if Pr ( substitution candidate | target word ) > threshold Model type Substitution detection rate False alarm rate No predicted substitutions %2.42 % Pr >= %2.77 % Pr >= %3.67 %

9 Combining Rote and Extrapolative Aim: Get n substitutions for a given word. Step 1: Use top n substitutions from rote. Step 2: If rote predicts k substitutions, k < n, –Then add top n – k substitutions from extrapolative. Intuition: rote is more accurate, so use when available. If not available, fall back on extrapolative.

10 Results from Combining Algorithms Model type Substitution detection rate False alarm rate Top %3.77 % Truncation24.69 %4.29 % Top %4.62 % Theoretical max %2.90 % Truncation = The first 2 to n-2 phonemes of a word – models false starts. [/K AE/ and /K AE N/ for /K AE N D IY/; none for “hate”] Theoretical max = use only those miscues the child actually made.

11 Conclusion Evaluated effect on substitution detection of –Two previously published algorithms –A combination of the two algorithms. Combined approach improved on current configuration (truncations) by –Reducing false alarms by 0.52% abs (12% rel) –Increasing miscue detection by 1.04% (4.2% rel) Take-home sound byte: Listening for specific reading mistakes can help detect them!