Preposition error correction using Graph Convolutional Networks

Slides:



Advertisements
Similar presentations
School of something FACULTY OF OTHER School of Computing FACULTY OF ENGINEERING Chunking: Shallow Parsing Eric Atwell, Language Research Group.
Advertisements

Machine Learning Approaches to the Analysis of Large Corpora : A Survey Xunlei Rose Hu and Eric Atwell University of Leeds.
Using Syntax to Disambiguate Explicit Discourse Connectives in Text Source: ACL-IJCNLP 2009 Author: Emily Pitler and Ani Nenkova Reporter: Yong-Xiang Chen.
Recap of parts of speech Parts of speech: Sexier than Ryan Gosling.
Part of Speech Tagging Importance Resolving ambiguities by assigning lower probabilities to words that don’t fit Applying to language grammatical rules.
Using Web Queries for Learner Error Detection Michael Gamon, Microsoft Research Claudia Leacock, Butler-Hill Group.
Inducing Information Extraction Systems for New Languages via Cross-Language Projection Ellen Riloff University of Utah Charles Schafer, David Yarowksy.
1 Combining Lexical and Syntactic Features for Supervised Word Sense Disambiguation Saif Mohammad Ted Pedersen University of Toronto University of Minnesota.
WEEK 6: DEEP TRACKING STUDENTS: SI CHEN & MEERA HAHN MENTOR: AFSHIN DEGHAN.
SVMLight SVMLight is an implementation of Support Vector Machine (SVM) in C. Download source from :
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Overview Project Goals –Represent a sentence in a parse tree –Use parses in tree to search another tree containing ontology of project management deliverables.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
Detection of Links between Words in the Task of Syntactic-Semantic Analysis of Russian Texts. Dmitry V. Merkuryev Saint-Petersburg State University, Russia.
The man bites the dog man bites the dog bites the dog the dog dog Parse Tree NP A N the man bites the dog V N NP S VP A 1. Sentence  noun-phrase verb-phrase.
Page 1 NAACL-HLT 2010 Los Angeles, CA Training Paradigms for Correcting Errors in Grammar and Usage Alla Rozovskaya and Dan Roth University of Illinois.
Deep Learning for Efficient Discriminative Parsing Niranjan Balasubramanian September 2 nd, 2015 Slides based on Ronan Collobert’s Paper and video from.
Semantic Compositionality through Recursive Matrix-Vector Spaces
11 Project, Part 3. Outline Basics of supervised learning using Naïve Bayes (using a simpler example) Features for the project 2.
ESL 11B Class structure and student requirements.
IN READING, LANGUAGE, AND LISTENING Grades 6-8 Florida Standards Assessment in English Language Arts.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
ACL/EMNLP 2012 review (eNLP version) Mamoru Komachi 2012/07/17 Educational NLP research group Computational Linguistics Lab Nara Institute of Science and.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
Learning to Generate Complex Morphology for Machine Translation Einat Minkov †, Kristina Toutanova* and Hisami Suzuki* *Microsoft Research † Carnegie Mellon.
This research is supported by the U.S. Department of Education and DARPA. Focuses on mistakes in determiner and preposition usage made by non-native speakers.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Bassem Makni SML 16 Click to add text 1 Deep Learning of RDF rules Semantic Machine Learning.
Natural Language Processing Vasile Rus
Automating rule generation for grammar checkers
Language Identification and Part-of-Speech Tagging
“Leiningen vs the Ants”
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
CSC 594 Topics in AI – Natural Language Processing
Google SyntaxNet “Parsey McParseface and other SyntaxNet models are some of the most complex networks that we have trained with the TensorFlow framework.
David Mareček and Zdeněk Žabokrtský
Relation Extraction CSCI-GA.2591
Grades 6-8 Florida Standards Assessment in English Language Arts
TeamMember1 TeamMember2 Machine Learning Project 2016/2017-2
Inception and Residual Architecture in Deep Convolutional Networks
Giuseppe Attardi Dipartimento di Informatica Università di Pisa
Bird-species Recognition Using Convolutional Neural Network
Quanzeng You, Jiebo Luo, Hailin Jin and Jianchao Yang
Lei Sha, Jing Liu, Chin-Yew Lin, Sujian Li, Baobao Chang, Zhifang Sui
web1T and deep learning methods
Recurrent Neural Networks
Extracting Semantic Concept Relations
Transformer result, convolutional encoder-decoder
Eiji Aramaki* Sadao Kurohashi* * University of Tokyo
Giuseppe Attardi Dipartimento di Informatica Università di Pisa
The CoNLL-2014 Shared Task on Grammatical Error Correction
Seminar Topics and Projects
Automatic Detection of Causal Relations for Question Answering
Age and Gender Classification using Convolutional Neural Networks
Lecture: Deep Convolutional Neural Networks
Tuning CNN: Tips & Tricks
Natural Language Processing
Giuseppe Attardi Dipartimento di Informatica Università di Pisa
Word embeddings (continued)
Attention for translation
CIS 519 Recitation 11/15/18.
Presented By: Sparsh Gupta Anmol Popli Hammad Abdullah Ayyubi
Word representations David Kauchak CS158 – Fall 2016.
Neural Machine Translation using CNN
Extracting Why Text Segment from Web Based on Grammar-gram
Presented By: Harshul Gupta
Baseline Model CSV Files Pandas DataFrame Sentence Lists
SUCCESS CRITERIA: UNDERSTAND BASIC SENTENCES FOR GRADE E.
CS249: Neural Language Model
Presentation transcript:

Preposition error correction using Graph Convolutional Networks Anant Gupta

Problem Preposition error correction ⊂ Grammar error correction The semester starts on January.                     We reached at the airport on time. The semester starts in January.                      We reached the airport on time. (Replacement)                                                                                     (Deletion) This is a comfortable house to live. This is a comfortable house to live in. (Insertion)

Problem Preposition error correction ⊂ Grammar error correction The semester starts on January.                     We reached at the airport on time. The semester starts in January.                      We reached the airport on time. (Replacement) - This project                                                               (Deletion) This is a comfortable house to live. This is a comfortable house to live in. (Insertion)

Approaches Classification - Articles, Prepositions, Noun number Machine Translation - Broad class of errors

Approaches Classification - Articles, Prepositions, Noun number Machine Translation - Broad class of errors In this project: Preposition classification

Idea Use context (rest of sentence) to predict preposition Dependency parse tree provides natural structure

Dependency parse tree I asked the professor for an extension

Idea Perform convolutions on parse tree Learn features (word -> phrase embeddings) Classify using features of neighbours

Idea Perform convolutions on parse tree Learn features (word -> phrase embeddings) Classify using features of neighbours

Idea Perform convolutions on parse tree Learn features (word -> phrase embeddings) Classify using features of neighbours

Network Architecture Pretrained word embeddings (GloVe)

Experiment details Training and eval data BAWE corpus Collection of assignment texts from universities, English discipline assignments 22,000 + 2,500 train-dev split Test data CoNLL-14 shared task test data 2650 correct examples, 107 erroneous examples

Experiment details Frameworks used SyntaxNet (Dependency parsing and POS tags) Tensorflow (for training)

Results Baseline (parent child word embeddings, fully connected layers): Evaluation accuracy: 65.78% Test accuracy: 50.20% Experiment (tree convolution): Evaluation accuracy: 66.56% Test accuracy: 54.15%

Plots Training and eval accuracy vs # training steps (Experiment)

Plots Training and eval accuracy vs # training steps (Baseline)

Future work Try higher dimension embeddings with more training data Look at other ways to learn phrase embeddings Identify which examples are “hard” (need more context than just parent and child)

References http://tcci.ccf.org.cn/conference/2015/papers/173.pdf https://arxiv.org/abs/1609.02907 https://arxiv.org/pdf/1606.00189.pdf http://www.aclweb.org/anthology/N16-1042 https://dl.acm.org/citation.cfm?id=2002589 http://www.comp.nus.edu.sg/~nlp/conll14st/CoNLLST01.pdf