Towards Learning Dialogue Structures from Speech Data and Domain Knowledge: Challenges to Conceptual Clustering using Multiple and Complex Knowledge Source.

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Chronos: A Tool for Handling Temporal Ontologies in Protégé
Albert Gatt Corpora and Statistical Methods Lecture 13.
1 Latent Semantic Mapping: Dimensionality Reduction via Globally Optimal Continuous Parameter Modeling Jerome R. Bellegarda.
Sequence Clustering and Labeling for Unsupervised Query Intent Discovery Speaker: Po-Hsien Shih Advisor: Jia-Ling Koh Source: WSDM’12 Date: 1 November,
NYU ANLP-00 1 Automatic Discovery of Scenario-Level Patterns for Information Extraction Roman Yangarber Ralph Grishman Pasi Tapanainen Silja Huttunen.
Sentiment Analysis An Overview of Concepts and Selected Techniques.
NLP and Speech Course Review. Morphological Analyzer Lexicon Part-of-Speech (POS) Tagging Grammar Rules Parser thethe – determiner Det NP → Det.
Predicting Text Quality for Scientific Articles Annie Louis University of Pennsylvania Advisor: Ani Nenkova.
Information Extraction CS 652 Information Extraction and Integration.
1 Lecture 5: Automatic cluster detection Lecture 6: Artificial neural networks Lecture 7: Evaluation of discovered knowledge Brief introduction to lectures.
Jierui Xie, Boleslaw Szymanski, Mohammed J. Zaki Department of Computer Science Rensselaer Polytechnic Institute Troy, NY 12180, USA {xiej2, szymansk,
Learning to Extract Form Labels Nguyen et al.. The Challenge We want to retrieve and integrate online databases We want to retrieve and integrate online.
School of Computing and Mathematics, University of Huddersfield Knowledge Engineering: Issues for the Planning Community Lee McCluskey Department of Computing.
Unsupervised Training and Clustering Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Designing clustering methods for ontology building: The Mo’K workbench Authors: Gilles Bisson, Claire Nédellec and Dolores Cañamero Presenter: Ovidiu Fortu.
Article by: Feiyu Xu, Daniela Kurz, Jakub Piskorski, Sven Schmeier Article Summary by Mark Vickers.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
Towards Semantic Web: An Attribute- Driven Algorithm to Identifying an Ontology Associated with a Given Web Page Dan Su Department of Computer Science.
Distributional Clustering of English Words Fernando Pereira- AT&T Bell Laboratories, 600 Naftali Tishby- Dept. of Computer Science, Hebrew University Lillian.
Language Modeling Frameworks for Information Retrieval John Lafferty School of Computer Science Carnegie Mellon University.
Ontology Learning and Population from Text: Algorithms, Evaluation and Applications Chapters Presented by Sole.
Feature Selection for Automatic Taxonomy Induction The Features Input: Two terms Output: A numeric score, or. Lexical-Syntactic Patterns Co-occurrence.
Semantic and phonetic automatic reconstruction of medical dictations STEFAN PETRIK, CHRISTINA DREXEL, LEO FESSLER, JEREMY JANCSARY, ALEXANDRA KLEIN,GERNOT.
Lecture 1, 7/21/2005Natural Language Processing1 CS60057 Speech &Natural Language Processing Autumn 2005 Lecture 1 21 July 2005.
CAREERS IN LINGUISTICS OUTSIDE OF ACADEMIA CAREERS IN INDUSTRY.
Empirical Methods in Information Extraction Claire Cardie Appeared in AI Magazine, 18:4, Summarized by Seong-Bae Park.
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
Interactive Dialogue Systems Professor Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh Pittsburgh,
7-Speech Recognition Speech Recognition Concepts
Natural Language Processing Introduction. 2 Natural Language Processing We’re going to study what goes into getting computers to perform useful and interesting.
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
Unsupervised learning of Natural languages Eitan Volsky Yasmine Meroz.
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
CS 4705 Lecture 19 Word Sense Disambiguation. Overview Selectional restriction based approaches Robust techniques –Machine Learning Supervised Unsupervised.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
October 2005CSA3180 NLP1 CSA3180 Natural Language Processing Introduction and Course Overview.
Generic Tasks by Ihab M. Amer Graduate Student Computer Science Dept. AUC, Cairo, Egypt.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
Compiler Construction (CS-636)
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
金聲玉振 Taiwan Univ. & Academia Sinica 1 Spoken Dialogue in Information Retrieval Jia-lin Shen Oct. 22, 1998.
Wei Lu, Hwee Tou Ng, Wee Sun Lee National University of Singapore
From Words to Senses: A Case Study of Subjectivity Recognition Author: Fangzhong Su & Katja Markert (University of Leeds, UK) Source: COLING 2008 Reporter:
Integrating Multiple Knowledge Sources For Improved Speech Understanding Sherif Abdou, Michael Scordilis Department of Electrical and Computer Engineering,
Using decision trees to build an a framework for multivariate time- series classification 1 Present By Xiayi Kuang.
MMM2005The Chinese University of Hong Kong MMM2005 The Chinese University of Hong Kong 1 Video Summarization Using Mutual Reinforcement Principle and Shot.
Overview of Statistical NLP IR Group Meeting March 7, 2006.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Semantic Interoperability in GIS N. L. Sarda Suman Somavarapu.
Sentiment Analysis Using Common- Sense and Context Information Basant Agarwal 1,2, Namita Mittal 2, Pooja Bansal 2, and Sonal Garg 2 1 Department of Computer.
Of 24 lecture 11: ontology – mediation, merging & aligning.
Ontology Engineering and Feature Construction for Predicting Friendship Links in the Live Journal Social Network Author:Vikas Bahirwani 、 Doina Caragea.
Semi-Supervised Clustering
Linguistic knowledge for Speech recognition
Natural Language Processing (NLP)
Statistical NLP: Lecture 9
Information Organization: Clustering
CS4705 Natural Language Processing
An Overview of Concepts and Selected Techniques
CS4705 Natural Language Processing
Text Categorization Berlin Chen 2003 Reference:
Natural Language Processing (NLP)
Chapter 10: Compilers and Language Translation
Statistical NLP : Lecture 9 Word Sense Disambiguation
Architecture Issue in the New Disciple System
Natural Language Processing (NLP)
Presentation transcript:

Towards Learning Dialogue Structures from Speech Data and Domain Knowledge: Challenges to Conceptual Clustering using Multiple and Complex Knowledge Source Jens-Uwe Moller Natural Language Systems Division, Dept. of Computer Science, Univ. of Hamburg

Overview Dialog modeling based on a set of units called dialog act Dialog acts from theory doesn’t fit with a specific domain Labeling dialog is time consuming and subjective  learn an application specific dialog acts from speech data using conceptual clustering

The learning task Learning dialog acts from turns Unsupervised classification (no prior definition of dialog acts is given) Hierarchy classification with inspectable classifying rules

Features Domain knowledge: structure of task, task knowledge represented by goals and plans Word recognizer: word hypotheses Prosodic data: Pause & Stress mark important unit Lexical semantics Syntax (less important in spoken dialog) Semantics (larger units of lexical semantics)

COWEB Symbolic machine learning algorithm Build a classification tree Distinction between subnodes are made from a function overall attribute Support probabilistic data Support multiple overlapping hierarchies (for ambiguous case) Can handle multiple entries of one attribute (e.g. stream of words)

COWEB (2) Learning from simultaneous events Learn from structure data: Conceptual Graphs. Learn case descriptions from terminological descriptions Subsumption = correclation criterion over structured data. e.g. subsumption of individuals to classes

Metrics for Measuring Domain Independence of Semantic Classes Andrew Pargellis, Eric Fosler-Lussier, Alexandros Potamianos, Chin-Hui Lee Dialogue Systems Research Dept., Bell Labs, Lucent Technologies Murray Hill, NJ, USA

Introduction Employ semantic classes (concepts) from another domain Need to identify domain-independent concepts base on comparison across domain Domain-independent concepts should occur in similar syntactic (lexical) contexts across domains

Comparing concepts across domains Concept-comparison method Concept-projection method

Concept-comparison method Find the similarity between all pairs of concepts across the two domains Two concepts are similar if their respective bigram contexts are similar Use left and right context bigram language models

Kullback-Leibler (KL) distance Compare how san francisco and newark are used in the Travel domain with how comedies and westerns are used in the Movie domain Distance between two concepts

Concept-projection method How well a single concept from one domain is represented in another domain. How the words comedies and westerns are used in both domains Useful for identifying the degree of domain- independence for a particular concept.

Result: Concept-comparison

Result: Concept-projection

Concept Example

Semi-Automatic Acquisition of Domain-Specific Semantic Structures Siu K.C., Meng H.M. Human-Computer Communications Laboratory Department of Systems Engineering and Engineering Management The Chinese University of Hong Kong

Grammar induction Use unannotated corpora Portable across domain & language Output grammar has reasonable coverage of within-domain data and reject out-of-domain data Amenable to interactive refinement by human Support optional injection of prior knowledge

Spatial clustering Use kullback-liebler distance. use left and right context. Consider word with pre-set minimum occurrence. (set to 5) use left and right context. Consider word w1, w2 (later be c1, c2) pair-wise for words that have a least pre-set minimum occurrence. (set to 5)

Temporal clustering Use Mutual Information (MI). N-highest MI pairs are clustered (N=5 in experiment)  Do spatial clustering and temporal clustering iteratively  Post-process by human

Automatic Concept identification In goal-oriented conversations Ananlada Chotimongkol and Alexander I. Rudnicky Language Technologies Institute Carnegie Mellon University

Concept identification First step towards the goal of automatically inferring domain ontologies Goal-oriented human-human conversation has a clear structure This structure can be used to automatically identify domain topics, e.g. dialog classfication

Clustering algorithm Hierarchical clustering Mutual information based Criterion=minimize the loss of average mutual information Kullback-Lierbler based Criterion=word pair with minimum distance

Evaluation metrics Reference concept from class-based n- gram model Cluster concept=majority concept Precision Recall Singularity score (SS) Quality score (QS)