The CoNLL-2014 Shared Task on Grammatical Error Correction

Slides:



Advertisements
Similar presentations
Automatic Timeline Generation from News Articles Josh Taylor and Jessica Jenkins.
Advertisements

The Chinese Room: Understanding and Correcting Machine Translation This work has been supported by NSF Grants IIS Solution: The Chinese Room Conclusions.
Spelling Correction for Search Engine Queries Bruno Martins, Mario J. Silva In Proceedings of EsTAL-04, España for Natural Language Processing Presenter:
© Paradigm Publishing, Inc Word 2010 Level 2 Unit 1Formatting and Customizing Documents Chapter 2Proofing Documents.
1 A Comparative Evaluation of Deep and Shallow Approaches to the Automatic Detection of Common Grammatical Errors Joachim Wagner, Jennifer Foster, and.
1 Developing Statistic-based and Rule-based Grammar Checkers for Chinese ESL Learners Howard Chen Department of English National Taiwan Normal University.
Using Web Queries for Learner Error Detection Michael Gamon, Microsoft Research Claudia Leacock, Butler-Hill Group.
Inducing Information Extraction Systems for New Languages via Cross-Language Projection Ellen Riloff University of Utah Charles Schafer, David Yarowksy.
Basic Scientific Writing in English Lecture 3 Professor Ralph Kirby Faculty of Life Sciences Extension 7323 Room B322.
Word Processing. ► This is using a computer for:  Writing  EditingTEXT  Printing  Used to write letters, books, memos and produce posters etc.  A.
Writing a Research Paper. Step 1: Define your topic.
Automated Essay Evaluation Martin Angert Rachel Drossman.
1 A study on automatically extracted keywords in text categorization Authors:Anette Hulth and Be´ata B. Megyesi From:ACL 2006 Reporter: 陳永祥 Date:2007/10/16.
Learner corpus analysis and error annotation Xiaofei Lu CALPER 2010 Summer Workshop July 13, 2010.
The CoNLL-2013 Shared Task on Grammatical Error Correction Hwee Tou Ng, Yuanbin Wu, and Christian Hadiwinoto 1 Siew.
Researcher affiliation extraction from homepages I. Nagy, R. Farkas, M. Jelasity University of Szeged, Hungary.
CSA2050: Introduction to Computational Linguistics Part of Speech (POS) Tagging II Transformation Based Tagging Brill (1995)
Tracking Language Development with Learner Corpora Xiaofei Lu CALPER 2010 Summer Workshop July 12, 2010.
S1: Chapter 1 Mathematical Models Dr J Frost Last modified: 6 th September 2015.
An Investigation of Statistical Machine Translation (Spanish to English) Raghav Bashyal.
Seminar in Applied Corpus Linguistics: Introduction APLNG 597A Xiaofei Lu August 26, 2009.
Error Correction: For Dummies? Ellen Pratt, PhD. UPR Mayaguez.
Computer Literacy for IC 3 Unit 2: Using Productivity Software Chapter 3: Formatting and Organizing Paragraphs and Documents © 2010 Pearson Education,
Computational linguistics A brief overview. Computational Linguistics might be considered as a synonym of automatic processing of natural language, since.
Corpus-based generation of suggestions for correcting student errors Paper presented at AsiaLex August 2009 Richard Watson Todd KMUTT ©2009 Richard Watson.
1 Introduction to Natural Language Processing ( ) Language Modeling (and the Noisy Channel) AI-lab
Auckland 2012Kilgarriff: NLP and Corpus Processing1 The contribution of NLP: corpus processing.
Page 1 NAACL-HLT 2010 Los Angeles, CA Training Paradigms for Correcting Errors in Grammar and Usage Alla Rozovskaya and Dan Roth University of Illinois.
IELTS Intensive Writing part two. IELTS Writing Two parts of ielts writing Part one writing about a Graph, chart, diagram Part two is an essay.
Error Analysis of Two Types of Grammar for the purpose of Automatic Rule Refinement Ariadna Font Llitjós, Katharina Probst, Jaime Carbonell Language Technologies.
CSA2050: Introduction to Computational Linguistics Part of Speech (POS) Tagging II Transformation Based Tagging Brill (1995)
Correcting Misuse of Verb Forms John Lee, Stephanie Seneff Computer Science and Artificial Intelligence Laboratory, MIT, Cambridge ACL 2008.
A Simple English-to-Punjabi Translation System By : Shailendra Singh.
FLAX Shaoqun Wu and Ian H. Witten Computer Science Department Waikato University New Zealand Utilizing lexical data from a web-derived.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Extending LanguageTool, a style and grammar checker Daniel Naber, Marcin Miłkowski 11:15 – 13:00 Friday, Workshop Notes v2.0, updated
Automating rule generation for grammar checkers
Language Identification and Part-of-Speech Tagging
Writing Inspirations, 2017 Aalto University
Tutorial 1 – Creating a Document
Approaches to Machine Translation
Learning Usage of English KWICly with WebLEAP/DSR
PRESENTED BY: PEAR A BHUIYAN
Computational and Statistical Methods for Corpus Analysis: Overview
Custom rules on subject verb agreement
LanguageTool - Part A David Ling.
ENGLISH TEST 45 Minutes – 75 Questions
Writing Inspirations, Spring 2016 Aalto University
CSCI 5832 Natural Language Processing
Topics in Linguistics ENG 331
web1T and deep learning methods
Transformer result, convolutional encoder-decoder
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
The CoNLL-2014 Shared Task on Grammatical Error Correction
Hong Kong English in Students’ Writing
Approaches to Machine Translation
Improved Word Alignments Using the Web as a Corpus
Grammar correction – Data collection interface
LanguageTool David Ling.
Statistical n-gram David ling.
Using GOLD to Tracking L2 Development
Applied Linguistics Chapter Four: Corpus Linguistics
CUTM 4012: Methods of Teaching English
Ngram frequency smooting
University of Illinois System in HOO Text Correction Shared Task
English project More detail and the data collection system
Preposition error correction using Graph Convolutional Networks
Introduction to Sentiment Analysis
Tri-gram + LanguageTool
Some preliminary results
Presentation transcript:

The CoNLL-2014 Shared Task on Grammatical Error Correction Rule-based/statistical parts of different teams and our next steps David Ling 21-11-2017

Contents Rule-based/statistical parts of different teams To do: Team CAMB, Team POST, Team UFC Different approaches for different errors (RB/LM/MT) To do: Annotated student scripts of HK students in xml Extracting problematic patterns from non-annotated scripts using language models (LM) Develop a rule-based system Ngram/Machine translation system

Grammar checking approaches Rule-based usually refer to handcrafted rules Low recall but high accuracy Statistical corpus-driven rules, usually ngram or parser features (parser may give incorrect tags if the sentence is problematic) Phase-based statistical machine translation Higher recall, but with lower accuracy Deep learning Machine translation Recent common approaches: Different methods for different error types

CoNLL-2014 approaches overview Many hybrid systems with rule- based (RB) LM: language model MT: machine translation SIMPLE hand-crafted rules are used as a preliminary Reasons: Some error types are more regular than others: subject verb agreement (SVA) Some are hard: wrong collocation eg. Heavy rain  Thick rain

Team CAMB (University of Cambridge) (1st) Rule-based + machine translation: Ngram corpus-driven rules Cambridge Learner Corpus, 6M words, with errors and corrections annotated, non-public Extract unigrams, bigrams, and trigrams as errors, which (1) marked as incorrect for > 90% of the times they occur and (2) appeared over 5 times Eg. “thick rain” is marked as problematic every time in the corpus Low recall, but high precision We do not have such huge annotated corpus, therefore an alternative is proposed later SMT part is skipped in this power point phase-based statistical machine translation (SMT) Input text Rule-based (RB) Output text

Team POST (Pohang University of Science and Technology)(4th) Language model + Hand-crafted rules (without SMT) Only a few simple hand-crafted rules Subject verb agreement: check if a singular noun, ‘this’, ‘it’, ‘one’, or gerund is in front of a singular verb (5 tokens ahead) Language model (LM): Noun number error (catcats) Input text Hand-crafted rules (RB): Insertion (only articles: a, an, the) Hand-crafted rules (RB): Subject verb agreement, preposition Ngram frequency (LM): Deletion, replacement (raiserise)

Team POST (4th) – ngram frequency for deletion, replacement A list of candidate pairs with different windows y (extracted from UNCLE) Use Google N-gram corpus frequency to give replacement or not Schematic example: (too to) with window (2;1): Sentence: I go too school by bus ‘I go too school’: 0 times vs ‘I go to school’: 5 times Replace ‘too’ by ‘to’ Window size of a pair is the one with the highest accuracy in training However, I think the neural network approach from LanguageTool is better (faster with less memory)

Team UFC (11th) - Universit´e de Franche-Comt´e Statistical parsers + rule-based Stanford Parser Provides POSTag and word dependency Example: Bills on ports and immigration were submitted by Senator Brownback, Republican of Kansas. Handcrafted-rules based on parsers Subject verb agreement verb form word form Statistical parsers Stanford Parser TreeTagger Input text

Conclusions Different methods for different error types Error with regular pattern  Simple handcraft rules: SVA (subject verb agreement) Vform (verb form: a study in 2010 [shown>showed] that) Mec (spelling, punctuation), ArtOrDet, Nn (noun number) Error depends on nearby words  Ngram frequency/LanguageTool neural network Prep (preposition) Vform, Wform (word form: The sense of [guilty>guilt] can) Wci (wrong collocation) Error depends on context  Machine translation UM (unclear meaning) Rloc- (redundancy)

Suggested next steps (1) An annotated corpus From HK students, manually Target: 1k marked essays Performance evaluation, training, and statistical findings also add values to literatures Suggest to follow the xml format of UNCLE (28 error types, see right): more data, can perform comparisons with others Save time and effort to build a new framework Develop an input interface: ref Xml format of UNCLE: Information of a correction: Start/end paragraph and character positions error type, correction

Suggested next steps (2) Extract error patterns from non-annotated scripts with low ngram probability, compare with good style corpus (Wiki or native speakers’ essays) Ngram with low probability are (1) Problematic, (2) Bad style, or (3) Correct but uncommon usage Non-annotated HK students’ essays, Target: 5k or more, in txt or word format We may try to group the extracted ngrams and make rules Example sentence from a HSMC student: “Hence , he explain lots of people upload a video to YouTube .” 3-gram frequency with Wiki corpus ' Hence , he ': 61, ' , he explain ': 1, ' he explain lots ': 0, ' explain lots of ': 1, ' lots of people ': 2383, ' of people upload ': 1, ' people upload a ': 0, ' upload a video ': 49, ' a video to ': 138, ' video to YouTube ':401, ‘ to YouTube . ': 105 2-gram frequency with Wiki corpus ' Hence , ': 3044, ' , he ': 365657, ' he explain ': 10, ' explain lots ': 0, ' lots of ': 3700, ' of people ': 21564, ' people upload ': 1, ' upload a ': 33, ' a video ': 6769, ' video to ': 744, ' to YouTube ': 55, ' YouTube \\. ': 338

Suggested next steps If we further have scripts from various countries, we may even compare to give the unique style/ problems of HK students Good essays (books/ wiki corpus) HK student essays Non-HK student essays HK style and mistakes Non-HK style and mistakes Compare to give unique styles and mistakes of HK

Suggested next steps (3) The rule-based part Similar to LanguageTool, allows English teachers to make handcrafted rules using xml A function to search for false alarms during deciding the handcrafted rules Assuming texts in Wikipedia are correct Try to search the corpus to see if there are any matched patterns with the designed rule This helps a lot in finding exceptions (actually LanguageTool has this feature online) 4) Language model and the machine translation part

Computing English Charles 1 month? Develop a script marking interface Collect and organize non-annotated student scripts 3 month? Develop a rule-based system Mark scripts using the interface Extract problematic ngrams & system update Test the rule-based system N month? Statistical and translation Analyze the problematic ngram Meeting with Charles? Time

End The CoNLL-2014 Shared Task on Grammatical Error Correction Google Books N-gram Corpus used as a Grammar Checker (EACL 2012) N-gram based Statistical Grammar Checker for Bangla and English Phrase-based Machine Translation is State-of-the-Art for Automatic Grammatical Error Correction, 2016 Conference on Empirical Methods in Natural Language Processing, pages 1546–1556