Putting ontology alignment in context: Usage scenarios, deployment and evaluation in a library case Antoine Isaac Henk Matthezing Lourens van der Meij.

Slides:



Advertisements
Similar presentations
Dr. Leo Obrst MITRE Information Semantics Information Discovery & Understanding Command & Control Center February 6, 2014February 6, 2014February 6, 2014.
Advertisements

Controlled Vocabularies in TELPlus Antoine ISAAC Vrije Universiteit Amsterdam EDLProject Workshop November 2007.
OAEI 2007: Library Track Results Antoine Isaac, Lourens van der Meij, Shenghui Wang, Henk Matthezing Claus Zinn, Stefan Schlobach, Frank van Harmelen Ontology.
STITCH final event KB July Agenda Brief presentation of STITCH main achievements Demo: annotation suggestion at KB The future use of STITCH results.
Advanced Information Systems Laboratory Department of Computer Science and Systems Engineering GI-DAYS MÜNSTER A software tool.
Copyright 2008 Tieto Corporation Database merge. Copyright 2008 Tieto Corporation Table of contents Please, do not remove this slide if you want to use.
Advanced Searching Engineering Village.
A Linguistic Approach for Semantic Web Service Discovery International Symposium on Management Intelligent Systems 2012 (IS-MiS 2012) July 13, 2012 Jordy.
Maurice Hermans.  Ontologies  Ontology Mapping  Research Question  String Similarities  Winkler Extension  Proposed Extension  Evaluation  Results.
Leveraging Your Taxonomy to Increase User Productivity MAIQuery and TM Navtree.
Selecting Preservation Strategies for Web Archives Stephan Strodl, Andreas Rauber Department of Software.
MEASUREMENT. Measurement “If you can’t measure it, you can’t manage it.” Bob Donath, Consultant.
A web-based repository service for vocabularies and alignments in the Cultural Heritage domain Lourens van der Meij Antoine Isaac Claus Zinn.
Using quantitative aspects of alignment generation for argumentation on mappings Antoine Isaac, Cassia Trojahn, Shenghui Wang, Paulo Quaresma Vrije Universteit.
Notes on ThoughtLab / Athena WP4 November 13, 2009 Antoine Isaac
PROMPT: Algorithm and Tool for Automated Ontology Merging and Alignment Natalya Fridman Noy and Mark A. Musen.
Aligning Thesauri for an integrated Access to Cultural Heritage Collections Antoine ISAAC (including slides by Frank van Harmelen) STITCH Project UDC Conference.
The Value of Usage Scenarios for Thesaurus Alignment in Cultural Heritage Context Antoine Isaac, Claus Zinn, Henk Matthezing, Lourens van der Meij, Stefan.
An Empirical Study of Instance-Based Ontology Mapping Antoine Isaac, Lourens van der Meij, Stefan Schlobach, Shenghui Wang funded by NWO Vrije.
Multi-Concept Alignment and Evaluation Shenghui Wang, Antoine Isaac, Lourens van der Meij, Stefan Schlobach Ontology Matching Workshop Oct. 11 th, 2007.
Vocabulary Matching for Book Indexing Suggestion in Linked Libraries – A Prototype Implementation & Evaluation Antoine Isaac, Dirk Kramer, Lourens van.
ISP 433/633 Week 6 IR Evaluation. Why Evaluate? Determine if the system is desirable Make comparative assessments.
SemanTic Interoperability To access Cultural Heritage Frank van Harmelen Henk Matthezing Peter Wittenburg Marjolein van Gendt Antoine Isaac Lourens van.
Science Inquiry Minds-on Hands-on.
New Ways of Mapping Knowledge Organization Systems Using a Semi-Automatic Matching- Procedure for Building Up Vocabulary Crosswalks Andreas Oskar Kempf.
OMAP: An Implemented Framework for Automatically Aligning OWL Ontologies SWAP, December, 2005 Raphaël Troncy, Umberto Straccia ISTI-CNR
Carlos Lamsfus. ISWDS 2005 Galway, November 7th 2005 CENTRO DE TECNOLOGÍAS DE INTERACCIÓN VISUAL Y COMUNICACIONES VISUAL INTERACTION AND COMMUNICATIONS.
Semantic Interoperability Jérôme Euzenat INRIA & LIG France Natasha Noy Stanford University USA.
AQUAINT Kickoff Meeting – December 2001 Integrating Robust Semantics, Event Detection, Information Fusion, and Summarization for Multimedia Question Answering.
Unit 2: Engineering Design Process
Evaluation of Quality of Learning Scenarios and Their Suitability to Particular Learners’ Profiles Assoc. Prof. Dr. Eugenijus Kurilovas, Vilnius University,
Query Relevance Feedback and Ontologies How to Make Queries Better.
Automatic Lexical Annotation Applied to the SCARLET Ontology Matcher Laura Po and Sonia Bergamaschi DII, University of Modena and Reggio Emilia, Italy.
Europeana and semantic alignment of vocabularies Antoine Isaac Jacco van Ossenbruggen, Victor de Boer, Jan Wielemaker, Guus Schreiber Europeana & Vrije.
Multilingual Information Exchange APAN, Bangkok 27 January 2005
ICS-FORTH January 11, Thesaurus Mapping Martin Doerr Foundation for Research and Technology - Hellas Institute of Computer Science Bath, UK, January.
OpenURL Link Resolvers 101
Assessing Quality for Integration Based Data M. Denk, W. Grossmann Institute for Scientific Computing.
Information Retrieval Evaluation and the Retrieval Process.
Theory and Application of Database Systems A Hybrid Approach for Extending Ontology from Text He Wei.
10/18/20151 Business Process Management and Semantic Technologies B. Ramamurthy.
Transparency in Searching and Choosing Peer Reviewers Doris DEKLEVA SMREKAR, M.Sc.Arch. Central Technological Library at the University of Ljubljana, Trg.
The KOS interoperability in aquatic science field through mapping processes Carmen Reverté Reverté Aquatic Ecosystems Documentation Center. IRTA. (Sant.
A Classification of Schema-based Matching Approaches Pavel Shvaiko Meaning Coordination and Negotiation Workshop, ISWC 8 th November 2004, Hiroshima, Japan.
Very Large Cross-lingual Resources at OAEI 2008 Laura Hollink Véronique Malaisé Vrije Universiteit Amsterdam.
Personalized Interaction With Semantic Information Portals Eric Schwarzkopf DFKI
CMPS 435 F08 These slides are designed to accompany Web Engineering: A Practitioner’s Approach (McGraw-Hill 2008) by Roger Pressman and David Lowe, copyright.
SKOS. Ontologies Metadata –Resources marked-up with descriptions of their content. No good unless everyone speaks the same language; Terminologies –Provide.
Thesauri usage in information retrieval systems: example of LISTA and ERIC database thesaurus Kristina Feldvari Departmant of Information Sciences, Faculty.
Instance-based mapping between thesauri and folksonomies Christian Wartena Rogier Brussee Telematica Instituut.
Friday, October 15 Objective: Students will be able to present information they have found on an engineering career. Bell Ringer: Quick Write: What is.
Working with Ontologies Introduction to DOGMA and related research.
Issues in Ontology-based Information integration By Zhan Cui, Dean Jones and Paul O’Brien.
Jane Reid, AMSc IRIC, QMUL, 30/10/01 1 Information seeking Information-seeking models Search strategies Search tactics.
1 Open Ontology Repository initiative - Planning Meeting - Thu Co-conveners: PeterYim, LeoObrst & MikeDean ref.:
Ontology Evaluation, Metrics, and Metadata in NCBO BioPortal Natasha Noy Stanford University.
Acceso a la información mediante exploración de sintagmas Anselmo Peñas, Julio Gonzalo y Felisa Verdejo Dpto. Lenguajes y Sistemas Informáticos UNED III.
KB subject prediction tool. STITCH final event KB subject prediction prototype Introduction Subject prediction is a special case of book reindexing What.
Controlled Vocabulary & Thesaurus Design Associative Relationships & Thesauri.
Extracting value from grey literature Processes and technologies for aggregating and analysing the hidden Big Data treasure of the organisations.
A System for Automatic Personalized Tracking of Scientific Literature on the Web Tzachi Perlstein Yael Nir.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Ontology Evaluation Outline Motivation Evaluation Criteria Evaluation Measures Evaluation Approaches.
Assessing SNOMED CT for Large Scale eHealth Deployments in the EU Workpackage 2- Building new Evidence Daniel Karlsson, Linköping University Stefan Schulz,
YourDataStories: Transparency and Corruption Fighting through Data Interlinking and Visual Exploration Georgios Petasis1, Anna Triantafillou2, Eric Karstens3.
Lecture 12: Data Wrangling
CS 430: Information Discovery
Actively Learning Ontology Matching via User Interaction
Business Process Management and Semantic Technologies
Presentation transcript:

Putting ontology alignment in context: Usage scenarios, deployment and evaluation in a library case Antoine Isaac Henk Matthezing Lourens van der Meij Stefan Schlobach Shenghui Wang Claus Zinn

Putting ontology alignment in context Introduction Alignment technology can help solving important problems heterogeneity of description resources But: What for, exactly? How useful can it be? Consensus: generation and evaluation of alignment have to take into account applications Problem: (relatively) not much investigation on alignment applications and their requirements

Putting ontology alignment in context Putting alignment into context: approach Focusing on application scenarios For a given scenario What are the expected meaning and use of alignments? How to use results of current alignment tools? How to fit evaluation to application’s success criteria? Testing two hypotheses For a same scenario, different evaluation strategies can bring different results For two scenarios, evaluation results can differ for a same alignment, even with the most appropriate strategies

Putting ontology alignment in context Agenda The KB application context Focus on two scenarios Thesaurus merging Book re-indexing OAEI 2007 Library track scenario-specific evaluation

Putting ontology alignment in context Our application context National Library of the Netherlands (KB) 2 main collections Each described (indexed) by its own thesaurus

Putting ontology alignment in context Usage scenarios for thesaurus alignment at KB Concept-based search Retrieving GTT-indexed books using Brinkman concepts Book re-indexing Indexing GTT-indexed books with Brinkman concepts Integration of one thesaurus into the other Inserting GTT elements into the Brinkman thesaurus Thesaurus merging Building a new thesaurus from GTT and Brinkman Free-text search matching user search terms to both GTT or Brinkman concepts Navigation browse the 2 collections through a merged version of the thesauri

Putting ontology alignment in context Agenda The KB application context Focus on two scenarios Thesaurus merging Book re-indexing OAEI 2007 Library track scenario-specific evaluation

Putting ontology alignment in context Thesaurus merging scenario Merge two vocabularies in a single, unified one Requirement: for two concepts, say whether a (thesaurus) semantic relation holds Broader (BT), narrower (NT), related (RT) Equivalence (EQ), if the two concepts share a same meaning and should be merged in a single one Similar to ontology engineering cases [Euzenat & Shvaiko, 2007]

Putting ontology alignment in context Deploying alignments for thesaurus merging De facto standard for alignment results (e1,e2,relation,measure) Problem: relation “ = “, rdfs:subClassOf or owl:equivalentClass Adaption has to be made Danger of overcommitment or loosening Problem: confidence/similarity measure Meaning? Weighted mappings have to be made crisp (e.g. by threshold)

Putting ontology alignment in context Thesaurus merging: evaluation method Alignments are evaluated in terms of individual mappings Does the mapping relation apply? Quite similar to classical ontology alignment evaluation Mappings can be assessed directly

Putting ontology alignment in context Thesaurus merging evaluation measures Correctness: proportion of proposed links that are correct Completeness: how many correct links were retrieved IR measures of precision and recall against a gold standard can be used Eventually semantic versions [Euzenat] Note: when no gold standard is present, other measures for completeness can be considered: coverage over a set of proposed alignments, for comparative evaluation of alignment tools coverage over the thesauri can be helpful for practitioners

Putting ontology alignment in context Agenda The KB application context Focus on two scenarios Thesaurus merging Book re-indexing OAEI 2007 Library track scenario-specific evaluation

Putting ontology alignment in context Book re-indexing scenario Scenario: re-annotation of GTT-indexed books by Brinkman concepts If one thesaurus is dropped, legacy data has to be indexed according to the other voc. Automatically or semi-automatically

Putting ontology alignment in context Book re-indexing requirements Requirement for a re-indexing function: converting sets of concepts to sets of concepts post-coordination: co-occurrence matters {G1=“History”, G2=“the Netherlands”} for GTT a book about Dutch history granularity of two vocabularies differ {B1=“Netherlands; History”} for Brinkman

Putting ontology alignment in context Semantic interpretation of re-indexing function One-to-one case: g1 can be converted to b1 if: Ideal case: b1 is semantically equivalent to g1 But b1 could also be more general than g1 Loss of information OK if b1 is the most specific subsumer of g1’s meaning Indexing specificity rule …

Putting ontology alignment in context Deploying alignments for book re-indexing Results of existing tools may need re-interpretation Unclear semantics of mapping relations and weights As for thesaurus merging Single concepts involved in mappings We need conversion of sets of concepts Only a few matching tools perform multi-concept mappings [Euzenat & Shvaiko]

Putting ontology alignment in context Deploying alignments for book re-indexing Solution: generate rules from 1-1 mappings “Sport” exactMatch “Sport” + “Sport” exactMatch “Sport practice” => “Sport” -> {“Sport”, “Sportpractice”} Several aggregation strategies are possible Firing rules for books Several strategies, e.g. fire a rule for a book if its index includes rule’s antecedent Merge results to produce new annotations

Putting ontology alignment in context Re-indexing evaluation We do not assess the mappings, nor even the rules We assess their application for book indexing More end-to-end General method: compare the annotations produced with the alignment with the correct ones

Putting ontology alignment in context Re-indexing evaluation measures Annotation level: measure correctness and completeness of the set of produced concepts Precision, Recall, Jaccard overlap (general distance) Notice: counting over annotated books, not rules or concepts Rules and concepts used more often are more important

Putting ontology alignment in context Re-indexing evaluation measures Book level: counting matched books Books for which there is one good annotation Minimal hint about users’ (dis)satisfaction

Putting ontology alignment in context Re-indexing: automatic evaluation There is a gold standard!

Putting ontology alignment in context Human evaluation vs. automatic evaluation Problems when considering application constraints Indexing variability Several indexers may make different choices Automatic evaluation compares with a specific one Evaluation variability Only one expert judgment is considered per book indexing assessment Evaluation set bias Dually-indexed books may present specific characteristics

Putting ontology alignment in context Re-indexing: manual evaluation Human expert assesses candidate indices: would have they chosen the same concepts? A “maybe” answer is now possible A slightly different perspective on evaluation criteria Acceptability of candidate indices

Putting ontology alignment in context Agenda The KB application context Focus on two scenarios Thesaurus merging Book re-indexing OAEI 2007 Library track scenario-specific evaluation

Putting ontology alignment in context Ontology Alignment Evaluation Initiative (OAEI) Apply and evaluate aligners on different tracks/cases Campaigns organized since 2004, and every year More tracks, more realistic tracks Better results of alignment tools Important for scientific community! OAEI 2007 Library track: KB thesauri Participants: Falcon, DSSim, Silas Mostly exactMatch-mappings

Putting ontology alignment in context Thesaurus merging evaluation No gold standard available Comparison with “reference” lexical alignment Manual assessment for a sample of “extra” mappings Coverage: proportion of good mappings found (participants + reference)

Putting ontology alignment in context Thesaurus merging: evaluation results Falcon performs well: closest to lexical reference DSSim and Ossewaarde add more to the lexical reference Ossewaarde adds less than DSSim, but additions are better

Putting ontology alignment in context Book re-indexing: automatic evaluation results

Putting ontology alignment in context Book re-indexing: manual evaluation results Research question: quality of candidate annotations Performances are consistently higher than for automatic evaluation

Putting ontology alignment in context Book re-indexing: manual evaluation results Research question: evaluation variability Krippendorff’s agreement coefficient (alpha): 0.62 Research question: indexing variability For dually indexed books, almost 20% of original indices are not even acceptable!

Putting ontology alignment in context Conclusions: observations Variety of scenarios requiring alignment There are common requirements, but also differences Leading to different success criteria and evaluation measures There is a gap with current alignment tools Deployment efforts are required Confirmation that different alignment strategies perform differently on different scenarios Choosing appropriate strategies

Putting ontology alignment in context Take-home message Just highlighting flaws? No, application-specific evaluation also helps to improve state-of-the-art alignment technology Simple but necessary things Evaluation is not free When done carefully, it brings many benefits

Putting ontology alignment in context Thanks!