Automatic Readability Evaluation Using a Neural Network Vivaek Shivakumar October 29, 2009.

Slides:



Advertisements
Similar presentations
Tracking L2 Lexical and Syntactic Development Xiaofei Lu CALPER 2010 Summer Workshop July 14, 2010.
Advertisements

EE3P BEng Final Year Project – 1 st meeting SLaTE – Speech and Language Technology in Education Martin Russell
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
UNIT-III By Mr. M. V. Nikum (B.E.I.T). Programming Language Lexical and Syntactic features of a programming Language are specified by its grammar Language:-
ASSESSING SEARCH TERM STRENGTH IN SPOKEN TERM DETECTION Amir Harati and Joseph Picone Institute for Signal and Information Processing, Temple University.
In Search of a More Probable Parse: Experiments with DOP* and the Penn Chinese Treebank Aaron Meyers Linguistics 490 Winter 2009.
CSC 380 Algorithm Project Presentation Spam Detection Algorithms Kyle McCombs Bridget Kelly.
Search Engines and Information Retrieval
SIMULATION. Simulation Definition of Simulation Simulation Methodology Proposing a New Experiment Considerations When Using Computer Models Types of Simulations.
Text-To-Speech System for Marathi Miss. Deepa V. Kadam Indian Institute of Technology, Bombay.
1.3 Executing Programs. How is Computer Code Transformed into an Executable? Interpreters Compilers Hybrid systems.
Lexiles A Way to Determine Text Complexity. How do we know if the reading selections we are offering are challenging our students?
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
Understanding Research Results
The What, the Why, and the How California Reading Association October 18, 2014.
Computational Methods to Vocalize Arabic Texts H. Safadi*, O. Al Dakkak** & N. Ghneim**
Measuring Hint Level in Open Cloze Questions Juan Pino, Maxine Eskenazi Language Technologies Institute Carnegie Mellon University International Florida.
Lemmatization Tagging LELA /20 Lemmatization Basic form of annotation involving identification of underlying lemmas (lexemes) of the words in.
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Building Effective Assessments. Agenda  Brief overview of Assess2Know content development  Assessment building pre-planning  Cognitive factors  Building.
What is Readability?  A characteristic of text documents..  “the sum total of all those elements within a given piece of printed material that affect.
Search Engines and Information Retrieval Chapter 1.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
Algorithmic Problem Solving CMSC 201 Adapted from slides by Marie desJardins (Spring 2015 Prof Chang version)
Learning to Predict Readability using Diverse Linguistic Features Rohit J. Kate 1 Xiaoqiang Luo 2 Siddharth Patwardhan 2 Martin Franz 2 Radu Florian 2.
Chapter 10: Compilers and Language Translation Invitation to Computer Science, Java Version, Third Edition.
A Comparison of Features for Automatic Readability Assessment Lijun Feng 1 Matt Huenerfauth 1 Martin Jansche 2 No´emie Elhadad 3 1 City University of New.
Which of the two appears simple to you? 1 2.
UNIT 4: Readability Index Measurement 1. What is Readability? The feature of plain language that makes it easy to read Or Describes the ease with which.
Indices Using Weighted Sums and Averages Readability Indices.
Research Topics CSC Parallel Computing & Compilers CSC 3990.
The Literacy Event Chapter 2. Qualities of Effective Teachers Think about the good teachers you have had or have observed. In your opinion, what made.
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved. 1.
1 CS 385 Fall 2006 Chapter 1 AI: Early History and Applications.
Towards the better software metrics tool motivation and the first experiences Gordana Rakić Zoran Budimac.
Methods for Automatic Evaluation of Sentence Extract Summaries * G.Ravindra +, N.Balakrishnan +, K.R.Ramakrishnan * Supercomputer Education & Research.
Number Sense Disambiguation Stuart Moore Supervised by: Anna Korhonen (Computer Lab)‏ Sabine Buchholz (Toshiba CRL)‏
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
The What, the Why, and the How California Reading Association October 23, 2015 This presentation can be found on SlideShare at
Evaluating 6 th Grade Literature By: Lorraine M. Carmona Torres Prof. E. Lugo ENGG 633 December 2 nd, 2010.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
STD Approach Two general approaches: word-based and phonetics-based Goal is to rapidly detect the presence of a term in a large audio corpus of heterogeneous.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
NATURAL LANGUAGE PROCESSING
Machine Learning Lecture 1: Intro + Decision Trees Moshe Koppel Slides adapted from Tom Mitchell and from Dan Roth.
Feature Assignment LBSC 878 February 22, 1999 Douglas W. Oard and Dagobert Soergel.
Chapter 1: Preliminaries Lecture # 2. Chapter 1: Preliminaries Reasons for Studying Concepts of Programming Languages Programming Domains Language Evaluation.
Assistant Instructor Nian K. Ghafoor Feb Definition of Proposal Proposal is a plan for master’s thesis or doctoral dissertation which provides the.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Trial Teaching Strategies: Linking Testing to Teaching Mary Beth Curtis Center for Special Education March 31, 2009.
IDENTIFYING GREAT TEACHERS THROUGH THEIR ONLINE PRESENCE Evanthia Faliagka, Maria Rigou, Spiros Sirmakessis.
LISTENING: QUESTIONS OF LEVEL FRANCISCO FUENTES NICOLAS VALENZUELA.
WP4 Models and Contents Quality Assessment
Big data classification using neural network
Developing an early warning system combined with dynamic LMS data
Text-To-Speech System for English
College of Engineering
Improving a Pipeline Architecture for Shallow Discourse Parsing
Dr. A .K. Bhattacharyya Professor EEI(NE Region), AAU, Jorhat
The Literacy Event Chapter 2 11/12/2018 9:52 PM
Writing Analytics Clayton Clemens Vive Kumar.
LING/C SC 581: Advanced Computational Linguistics
The Literacy Event Chapter 2 11/20/2018 1:07 AM
UNIT 3: READABILITY INDEX MEASUREMENT
NAACL-HLT 2010 June 5, 2010 Jee Eun Kim (HUFS) & Kong Joo Lee (CNU)
Chapter 10: Compilers and Language Translation
Artificial Intelligence 2004 Speech & Natural Language Processing
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Presentation transcript:

Automatic Readability Evaluation Using a Neural Network Vivaek Shivakumar October 29, 2009

Background and Purpose Readability – how difficult it is to read and comprehend a text – used in educational settings, grade-level reading evaluation Traditional readability formulas – Invented in 20 th Century before the computer age – Use primitive surface linguistic features – Still widely used, even in computer applications – e.g., Flesch-Kincaid Grade Lv.= – Dale-Chall Raw Score = %DW AvSL

Background and Purpose Real measure of readability factors in – Surface features (e.g., syllables per word, average sentence length) – Syntactic features (sentence structure, e.g., number of subordinate clauses) Parse tree size (e.g., Feng, 2009) – Semantic features (meanings, e.g., lexical density) – *Pragmatics (context) (out of project scope)

Background and Purpose Goal: create a model to give a more accurate score of readability of text using sophisticated techniques – Machine learning, e.g., neural networks: can be used to create a model using textual features as inputs Supervised – using state grade level standards assessment tests for training set

Development Neural Network – (still to be implemented) – Will be supervised – training set: reading passages from state and national grade level assessments – Grade levels “teach” the model to get more accurate – The neural network readability model should reflect the relationships between the different inputs that will be used

Development Criteria/Features to be used as inputs (possible) – Average word length in syllables – Average sentence length in words – Average sizes of sentence's parse/dependency trees – Lexical density (index based on frequency of words in text compared to in English in gen.) – common/uncommon words – Other syntactic features such as the presence of certain dependency types, etc.

Development Surface feature statistics (e.g. word/sentence lengths) and percentage of uncommon words* – Trivial to implement *not finished Parse/Dependency trees – Using Stanford Parser (or another if faster) – Output is analyzed from easy-to-read format Neural network – Not trivial to implement – bulk of development

Development Example of problem of working with natural language: syllable demarcation irregularities Implementation used to count syllables: – Each group of consecutive vowels (a,e,i,o,u) counts towards a syllable, with the following exceptions: Final -ES, -ED, and -E are not counted as syllables (besides -LE, which is). The letter “y” is a vowel unless it starts a word or follows another vowel. Any word of three letters or less counts as one syllable.

Preliminary Testing Evaluating three readability formulas vs. “actual” grade levels – same with dependency/parse tree sizes – Investigate whether there is a relationship, and if so how strong Texts used: same as for neural network training set – 92 texts at various grade levels

Results of Prelim. Testing

Analysis of Prelim. Results Dependency and Parse tree sizes are very closely linearly associated – Makes sense to only use one or the other in neural network – All of the three readability formulas show some association with grade level – surface features are useful but not alone – None are consistent – high deviation – all are unreliable

Expected Results Ideally, neural network learns to evaluate U.S. Grade level of a given text with a significantly greater accuracy and precision than the existing formulas do