Speaker: Yun-Nung Chen 陳縕儂 Advisor: Prof. Lin-Shan Lee 李琳山 National Taiwan University Automatic Key Term Extraction and Summarization from Spoken Course.

Slides:



Advertisements
Similar presentations
Machine Learning Approaches to the Analysis of Large Corpora : A Survey Xunlei Rose Hu and Eric Atwell University of Leeds.
Advertisements

A Human-Centered Computing Framework to Enable Personalized News Video Recommendation (Oh Jun-hyuk)
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Atomatic summarization of voic messages using lexical and prosodic features Koumpis and Renals Presented by Daniel Vassilev.
Punctuation Generation Inspired Linguistic Features For Mandarin Prosodic Boundary Prediction CHEN-YU CHIANG, YIH-RU WANG AND SIN-HORNG CHEN 2012 ICASSP.
Sub-Project I Prosody, Tones and Text-To-Speech Synthesis Sin-Horng Chen (PI), Chiu-yu Tseng (Co-PI), Yih-Ru Wang (Co-PI), Yuan-Fu Liao (Co-PI), Lin-shan.
Acoustic Model Adaptation Based On Pronunciation Variability Analysis For Non-Native Speech Recognition Yoo Rhee Oh, Jae Sam Yoon, and Hong Kook Kim Dept.
15.0 Utterance Verification and Keyword/Key Phrase Spotting References: 1. “Speech Recognition and Utterance Verification Based on a Generalized Confidence.
Automatic Prosodic Event Detection Using Acoustic, Lexical, and Syntactic Evidence Sankaranarayanan Ananthakrishnan, Shrikanth S. Narayanan IEEE 2007 Min-Hsuan.
A New Suffix Tree Similarity Measure for Document Clustering Hung Chim, Xiaotie Deng City University of Hong Kong WWW 2007 Session: Similarity Search April.
A Statistical Model for Domain- Independent Text Segmentation Masao Utiyama and Hitoshi Isahura Presentation by Matthew Waymost.
47 th Annual Meeting of the Association for Computational Linguistics and 4 th International Joint Conference on Natural Language Processing Of the AFNLP.
Predicting Text Quality for Scientific Articles AAAI/SIGART-11 Doctoral Consortium Annie Louis : Louis A. and Nenkova A Automatically.
Combining Prosodic and Text Features for Segmentation of Mandarin Broadcast News Gina-Anne Levow University of Chicago SIGHAN July 25, 2004.
Ensemble Learning: An Introduction
1 Noun Homograph Disambiguation Using Local Context in Large Text Corpora Marti A. Hearst Presented by: Heng Ji Mar. 29, 2004.
Approaches to automatic summarization Lecture 5. Types of summaries Extracts – Sentences from the original document are displayed together to form a summary.
Improved Tone Modeling for Mandarin Broadcast News Speech Recognition Xin Lei 1, Manhung Siu 2, Mei-Yuh Hwang 1, Mari Ostendorf 1, Tan Lee 3 1 SSLI Lab,
Introduction to Machine Learning Approach Lecture 5.
Query session guided multi- document summarization THESIS PRESENTATION BY TAL BAUMEL ADVISOR: PROF. MICHAEL ELHADAD.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
1 Bayesian Learning for Latent Semantic Analysis Jen-Tzung Chien, Meng-Sun Wu and Chia-Sheng Wu Presenter: Hsuan-Sheng Chiu.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
Machine Learning in Spoken Language Processing Lecture 21 Spoken Language Processing Prof. Andrew Rosenberg.
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Summary  The task of extractive speech summarization is to select a set of salient sentences from an original spoken document and concatenate them to.
No. 1 Classification and clustering methods by probabilistic latent semantic indexing model A Short Course at Tamkang University Taipei, Taiwan, R.O.C.,
Japanese Spontaneous Spoken Document Retrieval Using NMF-Based Topic Models Xinhui Hu, Hideki Kashioka, Ryosuke Isotani, and Satoshi Nakamura National.
Yun-Nung (Vivian) Chen, Yu Huang, Sheng-Yi Kong, Lin-Shan Lee National Taiwan University, Taiwan.
1 Sentence-extractive automatic speech summarization and evaluation techniques Makoto Hirohata, Yosuke Shinnaka, Koji Iwano, Sadaoki Furui Presented by.
1 Reference Julian Kupiec, Jan Pedersen, Francine Chen, “A Trainable Document Summarizer”, SIGIR’95 Seattle WA USA, Xiaodan Zhu, Gerald Penn, “Evaluation.
Combining multiple learners Usman Roshan. Bagging Randomly sample training data Determine classifier C i on sampled data Goto step 1 and repeat m times.
Today Ensemble Methods. Recap of the course. Classifier Fusion
1 SIGIR 2004 Web-page Classification through Summarization Dou Shen Zheng Chen * Qiang Yang Presentation : Yao-Min Huang Date : 09/15/2004.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
Automatic Cue-Based Dialogue Act Tagging Discourse & Dialogue CMSC November 3, 2006.
Date : 2013/03/18 Author : Jeffrey Pound, Alexander K. Hudek, Ihab F. Ilyas, Grant Weddell Source : CIKM’12 Speaker : Er-Gang Liu Advisor : Prof. Jia-Ling.
Recognizing Discourse Structure: Speech Discourse & Dialogue CMSC October 11, 2006.
1 Prosody-Based Automatic Segmentation of Speech into Sentences and Topics Elizabeth Shriberg Andreas Stolcke Speech Technology and Research Laboratory.
Alignment of Bilingual Named Entities in Parallel Corpora Using Statistical Model Chun-Jen Lee Jason S. Chang Thomas C. Chuang AMTA 2004.
Mutual-reinforcement document summarization using embedded graph based sentence clustering for storytelling Zhengchen Zhang , Shuzhi Sam Ge , Hongsheng.
National Taiwan University, Taiwan
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Hello, Who is Calling? Can Words Reveal the Social Nature of Conversations?
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
10.0 Latent Semantic Analysis for Linguistic Processing References : 1. “Exploiting Latent Semantic Information in Statistical Language Modeling”, Proceedings.
Compact Query Term Selection Using Topically Related Text Date : 2013/10/09 Source : SIGIR’13 Authors : K. Tamsin Maxwell, W. Bruce Croft Advisor : Dr.Jia-ling,
Date: 2013/4/1 Author: Jaime I. Lopez-Veyna, Victor J. Sosa-Sosa, Ivan Lopez-Arevalo Source: KEYS’12 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang KESOSD.
Text Categorization by Boosting Automatically Extracted Concepts Lijuan Cai and Tommas Hofmann Department of Computer Science, Brown University SIGIR 2003.
CONTEXTUAL SEARCH AND NAME DISAMBIGUATION IN USING GRAPHS EINAT MINKOV, WILLIAM W. COHEN, ANDREW Y. NG SIGIR’06 Date: 2008/7/17 Advisor: Dr. Koh,
LexPageRank: Prestige in Multi-Document Text Summarization Gunes Erkan, Dragomir R. Radev (EMNLP 2004)
Statistical Models for Automatic Speech Recognition Lukáš Burget.
Improving Music Genre Classification Using Collaborative Tagging Data Ling Chen, Phillip Wright *, Wolfgang Nejdl Leibniz University Hannover * Georgia.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
The P YTHY Summarization System: Microsoft Research at DUC 2007 Kristina Toutanova, Chris Brockett, Michael Gamon, Jagadeesh Jagarlamudi, Hisami Suzuki,
Maximum Entropy techniques for exploiting syntactic, semantic and collocational dependencies in Language Modeling Sanjeev Khudanpur, Jun Wu Center for.
Meta-Path-Based Ranking with Pseudo Relevance Feedback on Heterogeneous Graph for Citation Recommendation By: Xiaozhong Liu, Yingying Yu, Chun Guo, Yizhou.
Recent Paper of Md. Akmal Haidar Meeting before ICASSP 2013 報告者:郝柏翰 2013/05/23.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
Statistical Models for Automatic Speech Recognition
11.0 Spoken Document Understanding and Organization for User-content Interaction References: 1. “Spoken Document Understanding and Organization”, IEEE.
Hsien-Chin Lin, Chi-Yu Yang, Hung-Yi Lee, Lin-shan Lee
A Brief Review of Extractive Summarization Research
Presented by Nick Janus
Automatic Prosodic Event Detection
Presentation transcript:

Speaker: Yun-Nung Chen 陳縕儂 Advisor: Prof. Lin-Shan Lee 李琳山 National Taiwan University Automatic Key Term Extraction and Summarization from Spoken Course Lectures 課程錄音之自動關鍵用語擷取及摘要

Introduction 2 Master Defense, National Taiwan University Target: extract key terms and summaries from course lectures

Key TermSummary O Indexing and retrieval O The relations between key terms and segments of documents 3 Introduction Master Defense, National Taiwan University O Efficiently understand the document Related to document understanding and semantics from the document Both are “Information Extraction”

4 Master Defense, National Taiwan University

Definition O Key Term O Higher term frequency O Core content O Two types O Keyword O Ex. “ 語音 ” O Key phrase O Ex. “ 語言 模型 ” 5 Master Defense, National Taiwan University

Automatic Key Term Extraction 6 ▼ Original spoken documents Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal ASR trans Master Defense, National Taiwan University

Automatic Key Term Extraction 7 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal ASR trans Master Defense, National Taiwan University

Automatic Key Term Extraction 8 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal ASR trans Master Defense, National Taiwan University

Phrase Identification Automatic Key Term Extraction 9 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal First using branching entropy to identify phrases ASR trans Master Defense, National Taiwan University

Phrase Identification Key Term Extraction Automatic Key Term Extraction 10 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal Key terms entropy acoustic model : Then using learning methods to extract key terms by some features ASR trans Master Defense, National Taiwan University

Phrase Identification Key Term Extraction Automatic Key Term Extraction 11 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal Key terms entropy acoustic model : ASR trans Master Defense, National Taiwan University

Branching Entropy 12 O Inside the phrase hidden Markov model How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University

Branching Entropy 13 O Inside the phrase hidden Markov model How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University

Branching Entropy 14 hidden Markov model boundary Define branching entropy to decide possible boundary How to decide the boundary of a phrase? represent is can : : is of in : : O Inside the phrase O Boundary of the phrase Master Defense, National Taiwan University

Branching Entropy 15 hidden Markov model O Definition of Right Branching Entropy O Probability of x i given X O Right branching entropy for X X xixi How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University

Branching Entropy 16 hidden Markov model O Decision of Right Boundary O Find the right boundary located between X and x i where X boundary How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University

Branching Entropy 17 hidden Markov model How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University

Branching Entropy 18 hidden Markov model How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University

Branching Entropy 19 hidden Markov model How to decide the boundary of a phrase? represent is can : : is of in : : Master Defense, National Taiwan University boundary Using PAT tree to implement

Phrase Identification Key Term Extraction Automatic Key Term Extraction 20 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal Key terms entropy acoustic model : Extract prosodic, lexical, and semantic features for each candidate term ASR trans Master Defense, National Taiwan University

Feature Extraction 21 O Prosodic features O For each candidate term appearing at the first time Feature Name Feature Description Duration (I – IV) normalized duration (max, min, mean, range) Speaker tends to use longer duration to emphasize key terms using 4 values for duration of the term duration of phone “a” normalized by avg duration of phone “a” Master Defense, National Taiwan University

Feature Extraction 22 O Prosodic features O For each candidate term appearing at the first time Higher pitch may represent significant information Feature Name Feature Description Duration (I – IV) normalized duration (max, min, mean, range) Master Defense, National Taiwan University

Feature Extraction 23 O Prosodic features O For each candidate term appearing at the first time Higher pitch may represent significant information Feature Name Feature Description Duration (I – IV) normalized duration (max, min, mean, range) Pitch (I - IV) F0 (max, min, mean, range) Master Defense, National Taiwan University

Feature Extraction 24 O Prosodic features O For each candidate term appearing at the first time Higher energy emphasizes important information Feature Name Feature Description Duration (I – IV) normalized duration (max, min, mean, range) Pitch (I - IV) F0 (max, min, mean, range) Master Defense, National Taiwan University

Feature Extraction 25 O Prosodic features O For each candidate term appearing at the first time Higher energy emphasizes important information Feature Name Feature Description Duration (I – IV) normalized duration (max, min, mean, range) Pitch (I - IV) F0 (max, min, mean, range) Energy (I - IV) energy (max, min, mean, range) Master Defense, National Taiwan University

Feature Extraction 26 O Lexical features Feature NameFeature Description TFterm frequency IDFinverse document frequency TFIDFtf * idf PoSthe PoS tag Using some well-known lexical features for each candidate term Master Defense, National Taiwan University

Feature Extraction 27 O Semantic features O Probabilistic Latent Semantic Analysis (PLSA) O Latent Topic Probability Key terms tend to focus on limited topics D i : documents T k : latent topics t j : terms Master Defense, National Taiwan University

Feature Extraction 28 O Semantic features O Probabilistic Latent Semantic Analysis (PLSA) O Latent Topic Probability Feature NameFeature Description LTP (I - III) Latent Topic Probability (mean, variance, standard deviation) non-key term key term Key terms tend to focus on limited topics describe a probability distribution Master Defense, National Taiwan University

Feature Extraction 29 O Semantic features O Probabilistic Latent Semantic Analysis (PLSA) O Latent Topic Significance Within-topic to out-of-topic ratio Feature NameFeature Description LTP (I - III) Latent Topic Probability (mean, variance, standard deviation) non-key term key term Key terms tend to focus on limited topics within-topic freq. out-of-topic freq. Master Defense, National Taiwan University

Feature Extraction 30 O Semantic features O Probabilistic Latent Semantic Analysis (PLSA) O Latent Topic Significance Within-topic to out-of-topic ratio Feature NameFeature Description LTP (I - III) Latent Topic Probability (mean, variance, standard deviation) LTS (I - III) Latent Topic Significance (mean, variance, standard deviation) non-key term key term Key terms tend to focus on limited topics within-topic freq. out-of-topic freq. Master Defense, National Taiwan University

Feature Extraction 31 O Semantic features O Probabilistic Latent Semantic Analysis (PLSA) O Latent Topic Entropy Feature NameFeature Description LTP (I - III) Latent Topic Probability (mean, variance, standard deviation) LTS (I - III) Latent Topic Significance (mean, variance, standard deviation) non-key term key term Key terms tend to focus on limited topics Master Defense, National Taiwan University

Feature Extraction 32 O Semantic features O Probabilistic Latent Semantic Analysis (PLSA) O Latent Topic Entropy Feature NameFeature Description LTP (I - III) Latent Topic Probability (mean, variance, standard deviation) LTS (I - III) Latent Topic Significance (mean, variance, standard deviation) LTEterm entropy for latent topic non-key term key term Key terms tend to focus on limited topics Higher LTE Lower LTE Master Defense, National Taiwan University

Phrase Identification Key Term Extraction Automatic Key Term Extraction 33 Archive of spoken documents Branching Entropy Feature Extraction Learning Methods 1)AdaBoost 2)Neural Network ASR speech signal ASR trans Key terms entropy acoustic model : Using supervised approaches to extract key terms Master Defense, National Taiwan University

Learning Methods 34 O Adaptive Boosting (AdaBoost) O Neural Network Automatically adjust the weights of features to train a classifier Master Defense, National Taiwan University

Automatic Key Term Extraction 35 Master Defense, National Taiwan University

Experiments 36 O Corpus O NTU lecture corpus O Mandarin Chinese embedded by English words O Single speaker O 45.2 hours O ASR System O Bilingual AM with model adaptation [1] O LM with adaptation using random forests [2] Master Defense, National Taiwan University LanguageMandarinEnglishOverall Char Acc (%) [1] Ching-Feng Yeh, “Bilingual Code-Mixed Acoustic Modeling by Unit Mapping and Model Recovery,” Master Thesis, [2] Chao-Yu Huang, “Language Model Adaptation for Mandarin-English Code-Mixed Lectures Using Word Classes and Random Forests,” Master Thesis, 2011.

Experiments 37 O Reference Key Terms O Annotations from 61 students who have taken the course O If the an annotator labeled 150 key terms, he gave each of them a score of 1/150, but 0 to others O Rank the terms by the sum of all scores given by all annotators for each term O Choose the top N terms form the list O N is average number of key terms O N = 154 key terms O 59 key phrases and 95 keywords O Evaluation O 3-fold cross validation Master Defense, National Taiwan University

Experiments 38 O Feature Effectiveness O Neural network for keywords from ASR transcriptions Each set of these features alone gives F1 from 20% to 42% Prosodic features and lexical features are additiveThree sets of features are all useful Pr: Prosodic Lx: Lexical Sm: Semantic F-measure Master Defense, National Taiwan University

Experiments 39 O Overall Performance (Keywords & Key Phrases) Baseline Branching entropy performs well F-measure Master Defense, National Taiwan University N-Gram TFIDF Branching Entropy TFIDF Branching Entropy AdaBoost Branching Entropy Neural Network key phrase keyword

The performance of manual is slightly better than ASR Experiments 40 O Overall Performance (Keywords & Key Phrases) Baseline Supervised learning using neural network gives the best results F-measure Master Defense, National Taiwan University N-Gram TFIDF Branching Entropy TFIDF Branching Entropy AdaBoost Branching Entropy Neural Network key phrase keyword

41 Master Defense, National Taiwan University

Introduction 42 O Extractive Summary O Important sentences in the document O Computing Importance of Sentences O Statistical Measure, Linguistic Measure, Confidence Score, N-Gram Score, Grammatical Structure Score O Ranking Sentences by Importance and Deciding Ratio of Summary Master Defense, National Taiwan University Proposed a better statistical measure of a term

Statistical Measure of a Term 43 O LTE-Based Statistical Measure (Baseline) O Key-Term-Based Statistical Measure O Considering only key terms O Weighted by LTS of the term Master Defense, National Taiwan University T k-1 TkTk T k+1 … … Key terms can represent core content of the document Latent topic probability can be estimated more accurately t i key

Importance of the Sentence 44 O Original Importance O LTE-based statistical measure O Key-term-based statistical measure O New Importance O Considering original importance and similarity of other sentences Master Defense, National Taiwan University Sentences similar to more sentences should get higher importance

Random Walk on a Graph 45 O Idea O Sentences similar to more important sentences should be more important O Graph Construction O Node: sentence in the document O Edge: weighted by similarity between nodes O Node Score O Interpolating two scores O Normalized original score of sentence S i O Scores propagated from neighbors according to edge weight p(j, i) Master Defense, National Taiwan University Nodes connecting to more nodes with higher scores should get higher scores score of S i in k-th iter.

Random Walk on a Graph 46 O Topical Similarity between Sentences O Edge weight sim(S i, S j ): (sentence i  sentence j ) O Latent topic probability of the sentence O Using Latent Topic Significance Master Defense, National Taiwan University SjSj t LTS SiSi … … TkTk T k+1 tjtj T k-1 titi tktk

Random Walk on a Graph 47 O Scores of Sentences O Converged equation O Matrix form O Solution dominate eigen vector of P’ O Integrated with Original Importance Master Defense, National Taiwan University

Automatic Summarization 48 Master Defense, National Taiwan University

Experiments 49 O Same Corpus and ASR System O NTU lecture corpus O Reference Summaries O Two human produced reference summaries for each document O Ranking sentences from “the most important” to “of average importance” O Evaluation Metric O ROUGE-1, ROUGE-2, ROUGE-3 O ROUGE-L: Longest Common Subsequence (LCS) Master Defense, National Taiwan University

Evaluation 50 Master Defense, National Taiwan University ASR ROUGE-1 LTE Key ROUGE-2 ROUGE-3 ROUGE-L Key-term-based statistical measure is helpful

Evaluation 51 Master Defense, National Taiwan University ROUGE-1ROUGE-2 ROUGE-3 ROUGE-L Random walk can help the LTE-based statistical measure ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L Random walk can also help the key-term-based statistical measure LTE LTE + RW Key Key + RW ASR Topical similarity can compensate recognition errors

Evaluation 52 Master Defense, National Taiwan University ROUGE-1ROUGE-2 ROUGE-3 ROUGE-L ASR LTE LTE + RW Key Key + RW Manual Key-term-based statistical measure and random walk using topical similarity are useful for summarization

53 Master Defense, National Taiwan University

Automatic Key Term ExtractionAutomatic Summarization The performance can be improved by ▫ Key-term-based statistical measure ▫ Random walk with topical similarity  Compensating recognition errors  Giving higher scores to sentences topically similar to more important sentences  Considering all sentences in the document 54 The performance can be improved by ▫ Identifying phrases by branching entropy ▫ Prosodic, lexical, and semantic features together Conclusions Master Defense, National Taiwan University

Published Papers: [1] Yun-Nung Chen, Yu Huang, Sheng-Yi Kong, and Lin-Shan Lee, “Automatic Key Term Extraction from Spoken Course Lectures Using Branching Entropy and Prosodic/Semantic Features,” in Proceedings of SLT, [2] Yun-Nung Chen, Yu Huang, Ching-Feng Yeh, and Lin-Shan Lee, “Spoken Lecture Summarization by Random Walk over a Graph Constructed with Automatically Extracted Key Terms,” in Proceedings of InterSpeech, Master Defense, National Taiwan University