Some preliminary results

Slides:



Advertisements
Similar presentations
MINING FEATURE-OPINION PAIRS AND THEIR RELIABILITY SCORES FROM WEB OPINION SOURCES Presented by Sole A. Kamal, M. Abulaish, and T. Anwar International.
Advertisements

Decision Errors and Power
A method for unsupervised broad-coverage lexical error detection and correction 4th Workshop on Innovative Uses of NLP for Building Educational Applications.
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
1 A Comparative Evaluation of Deep and Shallow Approaches to the Automatic Detection of Common Grammatical Errors Joachim Wagner, Jennifer Foster, and.
Chapter 9: The Normal Distribution
1 Developing Statistic-based and Rule-based Grammar Checkers for Chinese ESL Learners Howard Chen Department of English National Taiwan Normal University.
MULTIPLE REGRESSION. OVERVIEW What Makes it Multiple? What Makes it Multiple? Additional Assumptions Additional Assumptions Methods of Entering Variables.
Midterm Review CS4705 Natural Language Processing.
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Edge detection. Edge Detection in Images Finding the contour of objects in a scene.
Chapter 5 DESCRIBING DATA WITH Z-SCORES AND THE NORMAL CURVE.
Unit 5 Data Analysis.
STAT 3130 Statistical Methods I Session 2 One Way Analysis of Variance (ANOVA)
GRAMMAR APPROACH By: Katherine Marzán Concepción EDUC 413 Prof. Evelyn Lugo.
In 2009, the mean mathematics score was 21 with a standard deviation of 5.3 for the ACT mathematics section. ReferenceReference Draw the normal curve in.
Individual values of X Frequency How many individuals   Distribution of a population.
Probability and Samples
Sentence Review. 1. Use hamburger fold. 2. Fold edges down to fold.
CpSc 810: Machine Learning Evaluation of Classifier.
Distributions of the Sample Mean
Psychology 290 – Lab 9 January Normal Distribution Standardization Z-scores.
11/23/2015Slide 1 Using a combination of tables and plots from SPSS plus spreadsheets from Excel, we will show the linkage between correlation and linear.
Basic Implementation and Evaluations Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Lecture 8: Feature matching CS6670: Computer Vision Noah Snavely.
1 Probability and Statistics Confidence Intervals.
Correcting Misuse of Verb Forms John Lee, Stephanie Seneff Computer Science and Artificial Intelligence Laboratory, MIT, Cambridge ACL 2008.
N-Gram Model Formulas Word sequences Chain rule of probability Bigram approximation N-gram approximation.
Lesson planning I’d like to start by looking at two questions : Why ? How ? Those two questions will be answered if we answer the following three questions.
DOWeR Detecting Outliers in Web Service Requests Master’s Presentation of Christian Blass.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Welcome to the KS2 SATs Presentation Aims of the meeting:   To inform you about the Year 6 SATs   To discuss ways in which you can support your child’s.
Q035 The class, along with their teacher Mr. Freeman, ____ that the heater starts working. The right choice is ... pray prays Pray is the plural form.
Year 6 SATs meeting for Parents
Survey research: Ungraded review questions
Measuring Monolinguality
Chapter 6: INVENTORY COSTING
LECTURE 3: DATABASE SEARCHING PRINCIPLES
New to Year 6.
Corpus Linguistics I ENG 617
Week 10 Chapter 16. Confidence Intervals for Proportions
Setting significance levels at the correct level
Year 6 SAT’s Information Meeting
Is a Positive Developmental-Behavioral Screening Score Sufficient to Justify Referral? A Review of Evidence and Theory  R. Christopher Sheldrick, PhD,
Quantitative design: Ungraded review questions
web1T and deep learning methods
Linear and Nonlinear Functions
Cypress Upper School Tuesday 2nd October 2018
Confidence Intervals for a Population Mean, Standard Deviation Known
Evaluating Classifiers (& other algorithms)
N-Gram Model Formulas Word sequences Chain rule of probability
The CoNLL-2014 Shared Task on Grammatical Error Correction
The CoNLL-2014 Shared Task on Grammatical Error Correction
Hong Kong English in Students’ Writing
Welcome to Anchorsholme Academy’s Year 6 SATs Information Meeting
Patricia Butterfield & Naomi Chaytor October 18th, 2017
Grammar correction – Data collection interface
Statistical n-gram David ling.
Chapter 8 Confidence Intervals.
Introduction to Gaussian Errors
Ngram frequency smooting
Sampling distributions:
A student attempts to calculate the missing length a.
Evaluating Classifiers
Teacher : ANGELA CASTIBLANCO
Retrieval Performance Evaluation - Measures
Quantitative design: Ungraded review questions
Tri-gram + LanguageTool
Precision and Recall Reminder:
Extracting Why Text Segment from Web Based on Grammar-gram
Presentation transcript:

Some preliminary results 2017-12-27

Marking interface

Highlight unusual phrases / grammatical error Current corpora: Google ngram, Wikipedia 2007 Too many phrases not in corpus Reasons for not presenting in the corpus: (1) uncommon words (2) small corpus (3) wrong usage Highlighted: trigram not in Google ngram book (from HSMC student’s report)

It seems that it is weak in identifying simple subject verb agreement To rule out false positive highlights, we attempted normalized linear approximation last week: It seems that it is weak in identifying simple subject verb agreement 𝑝′′′ 𝑤 1 , 𝑤 2 , 𝑤 3 = 𝜆 1 𝑝′ 𝑤 1 , 𝑤 2 , 𝑤 3 + 𝜆 2 𝑝′′ 𝑤 1 , 𝑤 2 , 𝑤 3 𝑠𝑐𝑜𝑟𝑒= 𝑝′′′ 𝑤 1 , 𝑤 2 , 𝑤 3 𝑝 𝑤 1 𝑝 𝑤 2 𝑝 𝑤 3

TEST 1 O: True Positive, X: False Positive (c) Normalized Linear approximation (threshold = 0.3e-24, weights = 0.5, 0.5) 1 ,X 2, O 3, O 1: FN 2: FN 4, O 3: FN 5, O 5: FN 6, O 6: FN 7, X 8, O 9,O 10, X 7: FN 8: FN Precision: 0.7 Recall: 7/((7+8) = 0.47

TEST 3, very poor Left column: system Right column: by teacher (c) Normalized Linear approximation (threshold = 0.3e-24, weights = 0.5, 0.5) 1, X 1, X 2, X X:19 3, O Precision: 1/3=0.3 Recall: 1/20 = 0.05

New scores (1) Normalized score without interpolation normalized raw frequency = 𝑓𝑟𝑒𝑞( 𝑠𝑜 ℎ𝑒 𝑑𝑜) 𝑓𝑟𝑒𝑞( 𝑠𝑜)𝑓𝑟𝑒𝑞( ℎ𝑒)𝑓𝑟𝑒𝑞( 𝑑𝑜) (2) Sore by ratio of inflected form Eg. “So he do not explain what is YouTube” Possible inflected forms: “so he does” Calculate the ratio of “so he does”/”so he do”

Sore by ratio of inflected form Step1: use parser to detect POS tag with 'VBZ','VBP', 'VB','VBD‘ Step2: screening using normalized raw frequency THRESHOLD IS LOWER THAN PURE NORMALIZED RAW FREQUENCY normalized raw frequency = 𝑓𝑟𝑒𝑞( 𝑠𝑜 ℎ𝑒 𝑑𝑜) 𝑓𝑟𝑒𝑞( 𝑠𝑜)𝑓𝑟𝑒𝑞( ℎ𝑒)𝑓𝑟𝑒𝑞( 𝑑𝑜) Step3: ratio Ratio = 𝑓𝑟𝑒𝑞(𝑠𝑜 ℎ𝑒 𝑑𝑜𝑒𝑠) 𝑓𝑟𝑒𝑞(𝑠𝑜 ℎ𝑒 𝑑𝑜) =21.78 Step4: highlight if higher than a threshold "so he do”:3487, “so he does”:75976 “so": 724571145, "he": 2055218371, "do": 558298911} “normalized raw frequency": 4.195374032065089e-24,

Pink: due to normalized raw frequency: (threshold = 0.5e-24) Purple: due to ratio of inflected form: (rawf_threshold, ratio_threshold = 1e-23, 5.5)