1/(19) GATE Evaluation Tools GATE Training Course October 2006 Kalina Bontcheva.

Slides:



Advertisements
Similar presentations
Slide 1 CheVi Tutorial Aniko Simon, Darryl Reid, Sing Yoong Khew, and Zsolt Zsoldos SimBioSys presents.
Advertisements

2017/3/25 Test Case Upgrade from “Test Case-Training Material v1.4.ppt” of Testing basics Authors: NganVK Version: 1.4 Last Update: Dec-2005.
User Guide TE Connectivity Data Collection Tool (DCT) January 2012.
©2011 1www.id-book.com Evaluation studies: From controlled to natural settings Chapter 14.
Slide 1 Insert your own content. Slide 2 Insert your own content.
1 End-User Programming to Support Classroom Activities on Small Devices Craig Prince University of Washington VL/HCC 2008.
eClassifier: Tool for Taxonomies
1 OOA-HR Workshop, 11 October 2006 Semantic Metadata Extraction using GATE Diana Maynard Natural Language Processing Group University of Sheffield, UK.
Multilinguality & Semantic Search Eelco Mossel (University of Hamburg) Review Meeting, January 2008, Zürich.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Introduction to HTML, XHTML, and CSS
0 - 0.
DIVIDING INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
SUBTRACTING INTEGERS 1. CHANGE THE SUBTRACTION SIGN TO ADDITION
MULT. INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
Addition Facts
Database Design: ER Modelling (Continued)
Corpus Linguistics Richard Xiao
ENV Envisioning Information Lecture 2 Simple Graphs and Charts Ken Brodlie School of Computing University of Leeds.
View-Based Application Development Lecture 1 1. Flows of Lecture 1 Before Lab Introduction to the Game to be developed in this workshop Comparison between.
1 NatQuery 3/05 An End-User Perspective On Using NatQuery To Extract Data From ADABAS Presented by Treehouse Software, Inc.
4. Internet Programming ENG224 INFORMATION TECHNOLOGY – Part I
LABELING TURKISH NEWS STORIES WITH CRF Prof. Dr. Eşref Adalı ISTANBUL TECHNICAL UNIVERSITY COMPUTER ENGINEERING 1.
BT Wholesale October Creating your own telephone network WHOLESALE CALLS LINE ASSOCIATED.
Report Card P Only 4 files are exported in SAMS, but there are at least 7 tables could be exported in WebSAMS. Report Card P contains 4 functions: Extract,
Addison Wesley is an imprint of © 2010 Pearson Addison-Wesley. All rights reserved. Chapter 10 Arrays and Tile Mapping Starting Out with Games & Graphics.
The Office Procedures and Technology
1 A Test Automation Tool For Java Applets Testing of Web Applications TATJA Program Demonstration Conclusions By Matthew Xuereb.
Campaign Overview Mailers Mailing Lists
PEPS Weekly Data Extracts User Guide September 2006.
Microsoft Access.
State of Connecticut Core-CT Project Query 8 hrs Updated 6/06/2006.
Ontology-based User Modeling for Web-based Information Systems Anton Andrejko, Michal Barla and Mária Bieliková {andrejko, barla,
1/(20) Introduction to ANNIE Diana Maynard University of Sheffield March 2004
An Introduction to GATE
26/10/2008 SWESE'08 1 Enhanced Semantic Access to Software Artefacts Danica Damljanović and Kalina Bontcheva.
Funded by: European Commission – 6th Framework Project Reference: IST WP6 review presentation GATE ontology QuestIO - Question-based Interface.
University of Sheffield, NLP Case study: GATE in the NeOn project Diana Maynard University of Sheffield.
Atomatic summarization of voic messages using lexical and prosodic features Koumpis and Renals Presented by Daniel Vassilev.
1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.
University of Sheffield NLP Exercise I Objective: Implement a ML component based on SVM to identify the following concepts in company profiles: company.
University of Sheffield NLP Module 4: Machine Learning.
1(18) GATE: A Unicode-based Infrastructure Supporting Multilingual Information Extraction Kalina Bontcheva, Diana Maynard, Valentin Tablan, Hamish Cunningham.
University of Sheffield NLP Module 11: Advanced Machine Learning.
Semantic multimedia annotation tool Tutorial authors : Batatia, Piombo
Addition 1’s to 20.
Dr. Alexandra I. Cristea XHTML.
25 seconds left…...
Test B, 100 Subtraction Facts
Week 1.
We will resume in: 25 Minutes.
CHAPTER 11 FILE INPUT & OUTPUT Introduction to Computer Science Using Ruby (c) 2012 Ophir Frieder et al.
The PLASTIC Model to HUTN transformation tool UDA.
On-line Compilation of Comparable Corpora and Their Evaluation Radu ION, Dan TUFIŞ, Tiberiu BOROŞ, Alexandru CEAUŞU and Dan ŞTEFĂNESCU Research Institute.
Chapter 8 Improving the User Interface
OU Campus User Training Tuesday, January 13, 2015.
CIS 4004: Web Based Information Technology Spring 2013
Jing-Shin Chang National Chi Nan University, IJCNLP-2013, Nagoya 2013/10/15 ACLCLP – Activities ( ) & Text Corpora.
University of Sheffield, NLP Annotation and Evaluation Diana Maynard, Niraj Aswani University of Sheffield.
Towards a semantic extraction of named entities Diana Maynard, Kalina Bontcheva, Hamish Cunningham University of Sheffield, UK.
INTRODUCTION TO ARTIFICIAL INTELLIGENCE Truc-Vien T. Nguyen Lab: Named Entity Recognition.
CROSSMARC Web Pages Collection: Crawling and Spidering Components Vangelis Karkaletsis Institute of Informatics & Telecommunications NCSR “Demokritos”
1/(13) Using Corpora and Evaluation Tools Diana Maynard Kalina Bontcheva
University of Sheffield, NLP Annotation and Evaluation Diana Maynard, Niraj Aswani University of Sheffield.
Introduction to GATE Developer Ian Roberts. University of Sheffield NLP Overview The GATE component model (CREOLE) Documents, annotations and corpora.
©2003 Paula Matuszek Taken primarily from a presentation by Lin Lin. CSC 9010: Text Mining Applications.
Exploiting Named Entity Taggers in a Second Language Thamar Solorio Computer Science Department National Institute of Astrophysics, Optics and Electronics.
Using Human Language Technology for Automatic Annotation and Indexing of Digital Library Content Kalina Bontcheva, Diana Maynard, Hamish Cunningham, Horacio.
Supervised Machine Learning
Social Knowledge Mining
Presentation transcript:

1/(19) GATE Evaluation Tools GATE Training Course October 2006 Kalina Bontcheva

2/(19) System development cycle 1.Collect corpus of texts 2.Annotate manually gold standard 3.Develop system 4.Evaluate performance 5.Go back to step 3, until desired performance is reached

3/(19) Corpora and System Development Gold standard data created by manual annotation Corpora are divided typically into a training and testing portion Rules and/or learning algorithms are developed or trained on the training part Tuned on the testing portion in order to optimise –Rule priorities, rules effectiveness, etc. –Parameters of the learning algorithm and the features used (typical routine: 10-fold cross validation) Evaluation set – the best system configuration is run on this data and the system performance is obtained No further tuning once evaluation set is used!

4/(19) Some NE Annotated Corpora MUC-6 and MUC-7 corpora - English CONLL shared task corpora - NEs in English and German - NEs in Spanish and Dutch TIDES surprise language exercise (NEs in Cebuano and Hindi) ACE – English

5/(19) Some NE Annotated Corpora MUC-6 and MUC-7 corpora - English CONLL shared task corpora - NEs in English and German - NEs in Spanish and Dutch TIDES surprise language exercise (NEs in Cebuano and Hindi) ACE – English

6/(19) The MUC-7 corpus 100 documents in SGML News domain Named Entities: 1880 Organizations (46%) 1324 Locations (32%) 887 Persons (22%) Inter-annotator agreement very high (~97%) roceedings/muc_7_proceedings/marsh_slides.pdfhttp:// roceedings/muc_7_proceedings/marsh_slides.pdf

7/(19) The MUC-7 Corpus (2) CAPE CANAVERAL, Fla. &MD; Working in chilly temperatures Wednesday night, NASA ground crews readied the space shuttle Endeavour for launch on a Japanese satellite retrieval mission. Endeavour, with an international crew of six, was set to blast off from the Kennedy Space Center on Thursday at 4:18 a.m. EST, the start of a 49-minute launching period. The nine day shuttle flight was to be the 12th launched in darkness.

8/(19) ACE – Towards Semantic Tagging of Entities MUC NE tags segments of text whenever that text represents the name of an entity In ACE (Automated Content Extraction), these names are viewed as mentions of the underlying entities. The main task is to detect (or infer) the mentions in the text of the entities themselves Rolls together the NE and CO tasks Domain- and genre-independent approaches ACE corpus contains newswire, broadcast news (ASR output and cleaned), and newspaper reports (OCR output and cleaned)

9/(19) ACE Entities Dealing with –Proper names – e.g., England, Mr. Smith, IBM –Pronouns – e.g., he, she, it –Nominal mentions – the company, the spokesman Identify which mentions in the text refer to which entities, e.g., –Tony Blair, Mr. Blair, he, the prime minister, he –Gordon Brown, he, Mr. Brown, the chancellor

10/(19) ACE Example <entity ID="ft-airlines-27-jul " GENERIC="FALSE" entity_type = "ORGANIZATION"> <entity_mention ID="M003" TYPE = "NAME" string = "National Air Traffic Services"> <entity_mention ID="M004" TYPE = "NAME" string = "NATS"> <entity_mention ID="M005" TYPE = "PRO" string = "its"> <entity_mention ID="M006" TYPE = "NAME" string = "Nats">

11/(19) Annotate Gold Standard – Manual Annotation in GATE GUI

12/(19) Ontology-Based Annotation (coming in GATE 4.0)

13/(19) Two GATE evaluation tools AnnotationDiff Corpus Benchmark Tool

14/(19) AnnotationDiff Graphical comparison of 2 sets of annotations Visual diff representation, like tkdiff Compares one document at a time, one annotation type at a time Gives scores for precision, recall, F_measure etc.

15/(19) Annotation Diff

16/(19) Corpus Benchmark Tool Compares annotations at the corpus level Compares all annotation types at the same time, i.e. gives an overall score, as well as a score for each annotation type Enables regression testing, i.e. comparison of 2 different versions against gold standard Visual display, can be exported to HTML Granularity of results: user can decide how much information to display Results in terms of Precision, Recall, F-measure

17/(19) Corpus structure Corpus benchmark tool requires particular directory structure Each corpus must have a clean and marked directory Clean holds the unannotated version, while marked holds the marked (gold standard) ones There may also be a processed subdirectory – this is a datastore (unlike the other two) Corresponding files in each subdirectory must have the same name

18/(19) How it works Clean, marked, and processed Corpus_tool.properties – must be in the directory from where gate is executed Specifies configuration information about –What annotation types are to be evaluated –Threshold below which to print out debug info –Input set name and key set name Modes –Default – regression testing –Human marked against already stored, processed –Human marked against current processing results

19/(19) Corpus Benchmark Tool