Summary  Extractive speech summarization aims to automatically select an indicative set of sentences from a spoken document to concisely represent the.

Slides:



Advertisements
Similar presentations
Document Summarization using Conditional Random Fields Dou Shen, Jian-Tao Sun, Hua Li, Qiang Yang, Zheng Chen IJCAI 2007 Hao-Chin Chang Department of Computer.
Advertisements

Punctuation Generation Inspired Linguistic Features For Mandarin Prosodic Boundary Prediction CHEN-YU CHIANG, YIH-RU WANG AND SIN-HORNG CHEN 2012 ICASSP.
1/1/ A Knowledge-based Approach to Citation Extraction Min-Yuh Day 1,2, Tzong-Han Tsai 1,3, Cheng-Lung Sung 1, Cheng-Wei Lee 1, Shih-Hung Wu 4, Chorng-Shyong.
Improved Neural Network Based Language Modelling and Adaptation J. Park, X. Liu, M.J.F. Gales and P.C. Woodland 2010 INTERSPEECH Bang-Xuan Huang Department.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Recurrent Neural Networks ECE 398BD Instructor: Shobha Vasudevan.
VARIABLE PRESELECTION LIST LENGTH ESTIMATION USING NEURAL NETWORKS IN A TELEPHONE SPEECH HYPOTHESIS-VERIFICATION SYSTEM J. Macías-Guarasa, J. Ferreiros,
Presented by Zeehasham Rasheed
1 USING CLASS WEIGHTING IN INTER-CLASS MLLR Sam-Joo Doh and Richard M. Stern Department of Electrical and Computer Engineering and School of Computer Science.
Application of RNNs to Language Processing Andrey Malinin, Shixiang Gu CUED Division F Speech Group.
Distributed Representations of Sentences and Documents
Image Denoising and Inpainting with Deep Neural Networks Junyuan Xie, Linli Xu, Enhong Chen School of Computer Science and Technology University of Science.
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
Multi-Style Language Model for Web Scale Information Retrieval Kuansan Wang, Xiaolong Li and Jianfeng Gao SIGIR 2010 Min-Hsuan Lai Department of Computer.
Lightly Supervised and Unsupervised Acoustic Model Training Lori Lamel, Jean-Luc Gauvain and Gilles Adda Spoken Language Processing Group, LIMSI, France.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
A Survey of ICASSP 2013 Language Model Department of Computer Science & Information Engineering National Taiwan Normal University 報告者:郝柏翰 2013/06/19.
1 Bayesian Learning for Latent Semantic Analysis Jen-Tzung Chien, Meng-Sun Wu and Chia-Sheng Wu Presenter: Hsuan-Sheng Chiu.
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
Modeling Documents by Combining Semantic Concepts with Unsupervised Statistical Learning Author: Chaitanya Chemudugunta America Holloway Padhraic Smyth.
Deep Learning Neural Network with Memory (1)
Summary  The task of extractive speech summarization is to select a set of salient sentences from an original spoken document and concatenate them to.
Regression Approaches to Voice Quality Control Based on One-to-Many Eigenvoice Conversion Kumi Ohta, Yamato Ohtani, Tomoki Toda, Hiroshi Saruwatari, and.
Chengjie Sun,Lei Lin, Yuan Chen, Bingquan Liu Harbin Institute of Technology School of Computer Science and Technology 1 19/11/ :09 PM.
Retrieval Models for Question and Answer Archives Xiaobing Xue, Jiwoon Jeon, W. Bruce Croft Computer Science Department University of Massachusetts, Google,
1 Sentence-extractive automatic speech summarization and evaluation techniques Makoto Hirohata, Yosuke Shinnaka, Koji Iwano, Sadaoki Furui Presented by.
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
LML Speech Recognition Speech Recognition Introduction I E.M. Bakker.
Bo Pang , Lillian Lee Department of Computer Science
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
Round-Robin Discrimination Model for Reranking ASR Hypotheses Takanobu Oba, Takaaki Hori, Atsushi Nakamura INTERSPEECH 2010 Min-Hsuan Lai Department of.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
A Model for Learning the Semantics of Pictures V. Lavrenko, R. Manmatha, J. Jeon Center for Intelligent Information Retrieval Computer Science Department,
Semantic v.s. Positions: Utilizing Balanced Proximity in Language Model Smoothing for Information Retrieval Rui Yan†, ♮, Han Jiang†, ♮, Mirella Lapata‡,
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Conditional Random Fields for ASR Jeremy Morris July 25, 2006.
Subproject II: Robustness in Speech Recognition. Members (1/2) Hsiao-Chuan Wang (PI) National Tsing Hua University Jeih-Weih Hung (Co-PI) National Chi.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Combining Speech Attributes for Speech Recognition Jeremy Morris November 9, 2006.
Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department.
Performance Comparison of Speaker and Emotion Recognition
MINIMUM WORD CLASSIFICATION ERROR TRAINING OF HMMS FOR AUTOMATIC SPEECH RECOGNITION Yueng-Tien, Lo Speech Lab, CSIE National.
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Presenter : Chien Shing Chen Author: Wei-Hao.
Relevance Language Modeling For Speech Recognition Kuan-Yu Chen and Berlin Chen National Taiwan Normal University, Taipei, Taiwan ICASSP /1/17.
Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.
A Novel Relational Learning-to- Rank Approach for Topic-focused Multi-Document Summarization Yadong Zhu, Yanyan Lan, Jiafeng Guo, Pan Du, Xueqi Cheng Institute.
Positional Language Modeling for Extractive Broadcast News Speech Summarization Shih-Hung Liu, Kuan-Yu Chen, Berlin Chen, Hsin-Min Wang, Hsu-Chun Yen,
St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences Recurrent Neural Network-based Language Modeling for an Automatic.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
A Maximum Entropy Language Model Integrating N-grams and Topic Dependencies for Conversational Speech Recognition Sanjeev Khudanpur and Jun Wu Johns Hopkins.
Spoken Language Group Chinese Information Processing Lab. Institute of Information Science Academia Sinica, Taipei, Taiwan
Maximum Entropy techniques for exploiting syntactic, semantic and collocational dependencies in Language Modeling Sanjeev Khudanpur, Jun Wu Center for.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
Automatically Labeled Data Generation for Large Scale Event Extraction
NSF Grant Number: IIS PI: Joseph Picone Institution: Mississippi State University Title: Integrating Prosody, Speech Recognition, Parsing In Spoken-Language.
Online Multiscale Dynamic Topic Models
Intelligent Information System Lab
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Background & Overview Proposed Model Experimental Results Future Work
An Improved Neural Network Algorithm for Classifying the Transmission Line Faults Slavko Vasilic Dr Mladen Kezunovic Texas A&M University.
Department of Electrical Engineering
Hierarchical Relational Models for Document Networks
Hsien-Chin Lin, Chi-Yu Yang, Hung-Yi Lee, Lin-shan Lee
Information Retrieval and Web Design
Da-Rong Liu, Kuan-Yu Chen, Hung-Yi Lee, Lin-shan Lee
Presentation transcript:

Summary  Extractive speech summarization aims to automatically select an indicative set of sentences from a spoken document to concisely represent the important aspects of the document  An emerging stream of work is to employ the language modeling framework in an unsupervised manner  The major challenge is how to formulate the sentence models and accurately estimate their parameters  We propose a novel and effective recurrent neural network language modeling (RNNLM) framework for speech summarization  The deduced models are able to render not only word usage cues but also long-span structural information of word co-occurrence relationships A Recurrent Neural Network Language Modeling Framework for Extractive Speech Summarization Kuan-Yu Chen †,*, Shih-Hung Liu †,*, Berlin Chen #, Hsin-Min Wang †, Wen-Lian Hsu †, Hsin-His Chen * † Institute of Information Science, Academia Sinica, Taiwan, # National Taiwan Normal University, Taiwan, * National Taiwan University, Taiwan The Proposed RNNLM Summarization Method Speech Recognition System Sentence S 1 Document- level RNNLM Sentence- Specific RNNLM Sentence Ranking Sentence S N … Speech Summary P RNNLM (D|S) Speech Signal Language Modeling Framework  A principal realization of using language modeling for summarization is to use a probabilistic generative paradigm for ranking each sentence S of a spoken document D to be summarized  The simplest way is to estimate a unigram language model (ULM) on the basis of the frequency of each distinct word w occurring in the sentence S, with the maximum likelihood (ML) criterion Recurrent Neural Network Language Modeling  RNNLM has recently emerged as a promising modeling framework for several tasks  The statistical cues of previously encountered word retained in the hidden layer can be fed back for predicting an arbitrary succeeding word  Both word usage cues and long-span structural information of word co-occurrence relationships can be take into account naturally RNNLM for Summarization  A hierarchical training strategy has been proposed to obtain the corresponding RNNLM model for each sentence 1)A document-level RNNLM model is trained for each document to be summarized The model will memorize both word usage and long- span word dependence cues in the document 2)The sentence-specific RNNLM model is trained based on the document-level RNNLM mode 3)The resulting sentence-specific RNNLM model can be used in place of, or to complement, the original sentence model (i.e., ULM) Experimental Results  Dataset: MATBN Corpus The corpus collected by the Academia Sinica and the Public Television Service Foundation of Taiwan between November 2001 and April 2003 TD: using the manual transcripts of spoken documents (without speech recognition errors) SD: using the speech recognition transcripts that may contain speech recognition errors  Baseline Approaches  Summarization results achieved by a few state-of-the-art unsupervised methods ULM shows competitive results when compared to the other state-of-the-art unsupervised methods, confirming the applicability of the language modeling approach for speech summarization  The RNNLM Framework  RNNLM: The deduced sentence-specific RNNLM model be used in isolation  RNNLM+ULM: The RNNLM model be linearly combined with the unigram language model ROUGE-1ROUGE-2ROUGE-L TD ULM BLM RNNLM RNNLM+ULM SD ULM BLM RNNLM RNNLM+ULM