Criterial features If you have examples of language use by learners (differentiated by L1 etc.) at different levels, you can use that to find the criterial.

Slides:



Advertisements
Similar presentations
You can use this presentation to: Gain an overall understanding of the purpose of the revised tool Learn about the changes that have been made Find advice.
Advertisements

Tracking L2 Lexical and Syntactic Development Xiaofei Lu CALPER 2010 Summer Workshop July 14, 2010.
Introduction to: Automated Essay Scoring (AES) Anat Ben-Simon Introduction to: Automated Essay Scoring (AES) Anat Ben-Simon National Institute for Testing.
Blueprints or conduits? Using an automated tool for text analysis Stuart G. Towns and Richard Watson Todd King Mongkut’s University of Technology Thonburi.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
Top Ten Tips for teachers preparing students for the academic version of IELTS Sam McCarter Macmillan Online Conference 2013.
Introduction.  Classification based on function role in classroom instruction  Placement assessment: administered at the beginning of instruction 
Using Rhetorical Grammar in the English 90 Classroom.
Automated Essay Evaluation Martin Angert Rachel Drossman.
A Demonstration Read-to-Write Project for Content Area Learning 2015 Summer Institute Joni Koehler and Dr. L. Lennie Irvin San Antonio Writing Project.
Chapter 7 Foregrounding Written Communication. Teaching Interactive Second Language Writing in Content- Based Classes Teachers should include a wide range.
What is Readability?  A characteristic of text documents..  “the sum total of all those elements within a given piece of printed material that affect.
CRESST ONR/NETC Meetings, July 2003, v1 ONR Advanced Distributed Learning Linguistic Modification of Test Items Jamal Abedi University of California,
What you have learned and how you can use it : Grammars and Lexicons Parts I-III.
The University of Illinois System in the CoNLL-2013 Shared Task Alla RozovskayaKai-Wei ChangMark SammonsDan Roth Cognitive Computation Group University.
Using Parallel Corpora for Contrastive Studies Michael Barlow.
Learning Targets We will understand the format and content of SpringBoard Close Reading Workshops. I will learn and apply strategies for close reading.
Language Identification and Part-of-Speech Tagging
Preparing for the College Writing Examination
Cultural Conversations
Automatic Writing Evaluation
ACT English Test Preparation
In Concert: An Integrated Reading and Writing Approach by Kathleen T
What do these mean? Your time is up Ready for anything (Red E)
E303 Part II The Context of Language Research
Vocabulary Module 2 Activity 5.
Intervention Strategies
Lexical bundles Last week we looked for ngrams
1. Review of last Friday (Form, Function, Fluency)
CSC 594 Topics in AI – Natural Language Processing
CORPUS LINGUISTICS Corpus linguistics is the study of language as expressed in samples (corpora) or "real world" text. An approach to derive at a set of.
Getting the Most from Writing
Anik Wulyani, PhD candidate
INAGO Project Automatic Knowledge Base Generation from Text for Interactive Question Answering.
Computational and Statistical Methods for Corpus Analysis: Overview
Content analysis, thematic analysis and grounded theory
Natural Language Processing (NLP)
Informational Writing Unit Grade 7-Looking at Grade 6/year 1
Learning About Language Assessment. Albany: Heinle & Heinle
Advanced English 6 November 1-2, 2017
CSCI 5832 Natural Language Processing
The key to your first draft [Outlines.pptx]
Writing Analytics Clayton Clemens Vive Kumar.
Topics in Linguistics ENG 331
Getting the Most from Writing
Teaching Writing LLT 307.
Machine Learning in Practice Lecture 11
HCC class lecture 13 comments
Statistical n-gram David ling.
Writing Learning Outcomes
GEE’S Writing RULES.
Writing Essays.
Using GOLD to Tracking L2 Development
Introduction to Text Analysis
Applied Linguistics Chapter Four: Corpus Linguistics
Lecture 6: How to Read an Academic Paper
Natural Language Processing (NLP)
Giannis Varelas Epimenidis Voutsakis Paraskevi Raftopoulou
Writing Instruction Kim Bowen, NCDPI ELA Consultant
Information Retrieval
The SAT Writing and Language Test
Writing the LEQ (Long Essay Question)
AP U.S. History Exam Details
Extracting Why Text Segment from Web Based on Grammar-gram
FCE IES Parque de Lisboa.
Using ELA Non-Summative Assessments for ELs Teacher Leader Summit 2019 Alice Garcia Special Population Assessment Coordinator.
The BAWE Quicklinks project
Natural Language Processing (NLP)
Presentation transcript:

Criterial features If you have examples of language use by learners (differentiated by L1 etc.) at different levels, you can use that to find the criterial features associated with each level of proficiency Those criterial features can be used in two ways: As a measure of expected proficiency and as a guide to the focus of language teaching As input to an automatic testing of the level of learners. (We return to this topic in a later class)

Criterial features There are different ways you can gather criterial features – for example: Lexicogrammatical features – perhaps using an n-gram analysis A large set of grammatical features. You could count the following sorts of features: passives, relative pronouns, participial clauses, nominalisations A grammatical analysis -- parsing

Criterial features The reading opts for the latter – note that this differs from the lexico-grammatical view I have described Step 1 is to tag the words with part-of-speech tags. There are about 50 or so POS tags, not just V, N, P, etc. E.g., CLAWS7 see http://ucrel.lancs.ac.uk/claws7tags.html

Criterial features Next stage is parsing – analysing sentences in terms of verb frames NP V PP (Noun Phrase – Verb – Prepostional Phrase) the cat sat on the mat Parsing in this case is based on a probabilistic grammar and so is corpus-based In addition, the grammatical relations are extracted – thus the cat above is identified as SUBJECT

Analysis of L2 language

Criterial features Having done this analysis, it is possible to analyse and compare students at different levels of proficiency, as discussed in the paper

Automatic Writing Evaluation

AWE Why might it be impossible? Why might it be possible?

AWE Why is it needed Use for testing and for feedback Shift from product to process approaches to writing Use of editing software (Word) and hence multiple drafts rather than typewriters (one or two drafts) Writing as a way to demonstrate proficiency Use for testing and for feedback

Warschauer & Page

Folz, Lochbaum & Rosenstein (Pearson)

What kinds of features potentially distinguish writing proficiency levels Ideas from previous classes and readings. ??

Samples Levels 1 to 5 Can you assess them correctly Are there features that can be used in an automatic system

New tools We will probably need to make use of a variety of tools and types of analysis Basic methodology create a large database of manually evaluated writing samples categorized by relevant features type of writing set prompts L1 (possibly) Discipline Concept – patterns can be extracted from large corpora (patterns may not be identifiable by humans)

General resources for language analysis Taggers (POS) Parsers/Tree taggers Semantic analysis Wordnet Framenet Semantic tagging

Language resources We can consult these resources manually In computer-based systems, they can be accessed automatically

Wordnet

Wordnet

Hypernym

Another search

Wordnet Here we checked the webpages Computational systems access Wordnet directly Use levels and “distances” (in tree structure) to assess different kinds and degrees of semantic relatedness

Framenet

Search for give

Semantic Tags Different ways of attempting semantic tags Give “sense” definitions of a word - run a company run [10] – sense 10 in COBUILD dictionary Provide semantic type information

Semantic tags Similar to a thesaurus organisation

Semantic tagging

Semantic tag list

Semantic tags Examples doubted_A7- UNLIKELY vague_A4.2- GENERAL best_A5.1+++ policy_X7+ EVALUATION GOOD WANTED truth_A5.2+ EVALUATION TRUE lies_A5.2- EVALUATION FALSE

AWE Features Different kinds of features and methods are used in AWE Surface features: word structure, lexical items, POS, sentence structure (latter two require tagging and parsing and so not overtly “surface”) Typically scores are higher with more words longer words subordination passives nominalizations prepositions

More sophisticated analyses Cohesion analysis Latent Semantic Analysis (LSA) Move analysis

AWE features Cohesive features Crossley and McNamara 2012. J. of Research in Reading 35, 2, 115-135 Lexical diversity – more word “types” Word overlap – to increase cohesion Synonyms Connectives

Coh-Metrix

Coh-Metrix

Crossley and McNamara 2012

Latent Semantic Analysis Large corpora are used to assess meaning similarity among words or passages The meaning of a word relates to all the passages in which the word appears, which can be represented as a matrix matrix (vector) for bus will more similar to car than biscuit

Latent Semantic Analysis The meaning of a passage relates to all the words that appear in the passage These vectors are the input to the processing that is something like a factor analysis Number of dimensions are reduced to give a more abstract and deeper analysis of word-context relations (maybe similar to human cognition)

Latent Semantic Analysis Initial analysis -- LSA space created by instructional materials or something similar plus student essays Different methods Some essays are graded by instructors and test essays are compared using LSA with these Exemplary essay is written and closeness of the test essays are judged Distance between each sentence of a standard text and each sentence of the student essay. A cumulative score is given All the student essays are compared and then ranked

Latent Semantic Analysis Depth is based on word choice and content Not syntax etc Students choose from a pre-existing set of prompts Typically requires a set of texts against which the test essays are judged

Move analysis Elena Cotos project at Iowa State Database of articles in different disciplines Examined “moves” in different sections of papers Created Intelligent Academic Discourse Evaluator (IADE) which is being used across the university

What is a Move? Move is a term used by Swales, referring to the function of each rhetorical unit (Swales 1990).

Using CorpusTools software Annotate or tag a text using an annotation scheme you devise: Examples For Errors For Evaluative terms For Moves

Corpus and layers Layer Corpus

Create a tagging scheme

Corpus and layers

Assign Tags to each unit Move: describe characteristics of cases/informants

Search Instances Tag Instances

kristopherkyle.com

AWE For testing For feedback what are the aims of an AWE system what kind of feedback should be given

Warschauer & Page

E-rater Identifies discourse elements: background, thesis, main ideas, supporting ideas, and conclusion – leads to a “development” score and a length score (number of words) Lexical complexity type/token ratio vocabulary level average word length

E-rater Prompt-specific vocabulary – based on content analysis (vector analysis comparing the words of the essay with the reference corpus) Essay length (number of words)

Weighting features All the assessment procedures are based on a sets of features – average word length, pronoun use, passive, … Statistical analyses will allow a weighting of each feature to produce a system that gives the best results – the most accurate scoring of the essays Features for feedback purposes are likely to be different – (Write more!)

Intelligent Essay Assessor (Pearson)

Pearson System

Holt Online Essay Scoring

Move feedback IADE – Elena Cotos

Move-annotated corpus: 20 articles × 50 disciplines = 1000 IADE is based on a move-annotated corpus.

Main Functions of IADE Look for DEFINITIONS of each step Check STEP STATISTICS Search ANNOTATED CORPUS Get REVISION TIPS

DEFINITIONS of each step: Examples (left), definition (right)

STEP STATISTICS

ANNOTATED CORPUS

REVISION TIPS

Criterion A commercial AWE system from ETS http://www.youtube.com/watch?v=62mft8hulDQ

AWE What is writing? What is the relation of AWE and learning?