1 Sentence-extractive automatic speech summarization and evaluation techniques Makoto Hirohata, Yosuke Shinnaka, Koji Iwano, Sadaoki Furui Presented by.

Slides:



Advertisements
Similar presentations
PHONE MODELING AND COMBINING DISCRIMINATIVE TRAINING FOR MANDARIN-ENGLISH BILINGUAL SPEECH RECOGNITION Yanmin Qian, Jia Liu ICASSP2010 Pei-Ning Chen CSIE.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
1 Latent Semantic Mapping: Dimensionality Reduction via Globally Optimal Continuous Parameter Modeling Jerome R. Bellegarda.
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
Chinese Word Segmentation Method for Domain-Special Machine Translation Su Chen; Zhang Yujie; Guo Zhen; Xu Jin’an Beijing Jiaotong University.
Latent Semantic Analysis
Confidence Estimation for Machine Translation J. Blatz et.al, Coling 04 SSLI MTRG 11/17/2004 Takahiro Shinozaki.
1 I256: Applied Natural Language Processing Marti Hearst Sept 27, 2006.
Presented by Zeehasham Rasheed
Name Extraction from Chinese Novels CS224n Spring 2008 Jing Chen and Raylene Yung.
Latent Semantic Analysis (LSA). Introduction to LSA Learning Model Uses Singular Value Decomposition (SVD) to simulate human learning of word and passage.
An Automatic Segmentation Method Combined with Length Descending and String Frequency Statistics for Chinese Shaohua Jiang, Yanzhong Dang Institute of.
Acoustic and Linguistic Characterization of Spontaneous Speech Masanobu Nakamura, Koji Iwano, and Sadaoki Furui Department of Computer Science Tokyo Institute.
Mining and Summarizing Customer Reviews
WebPage Summarization Using Clickthrough Data JianTao Sun & Yuchang Lu, TsingHua University, China Dou Shen & Qiang Yang, HK University of Science & Technology.
A VOICE ACTIVITY DETECTOR USING THE CHI-SQUARE TEST
RuleML-2007, Orlando, Florida1 Towards Knowledge Extraction from Weblogs and Rule-based Semantic Querying Xi Bai, Jigui Sun, Haiyan Che, Jin.
1 Bayesian Learning for Latent Semantic Analysis Jen-Tzung Chien, Meng-Sun Wu and Chia-Sheng Wu Presenter: Hsuan-Sheng Chiu.
Improving Utterance Verification Using a Smoothed Na ï ve Bayes Model Reporter : CHEN, TZAN HWEI Author :Alberto Sanchis, Alfons Juan and Enrique Vidal.
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
Summary  The task of extractive speech summarization is to select a set of salient sentences from an original spoken document and concatenate them to.
Classification and Ranking Approaches to Discriminative Language Modeling for ASR Erinç Dikici, Murat Semerci, Murat Saraçlar, Ethem Alpaydın 報告者:郝柏翰 2013/01/28.
Japanese Spontaneous Spoken Document Retrieval Using NMF-Based Topic Models Xinhui Hu, Hideki Kashioka, Ryosuke Isotani, and Satoshi Nakamura National.
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
1 Reference Julian Kupiec, Jan Pedersen, Francine Chen, “A Trainable Document Summarizer”, SIGIR’95 Seattle WA USA, Xiaodan Zhu, Gerald Penn, “Evaluation.
A Bootstrapping Method for Building Subjectivity Lexicons for Languages with Scarce Resources Author: Carmen Banea, Rada Mihalcea, Janyce Wiebe Source:
Luis Fernando D’Haro, Ondřej Glembek, Oldřich Plchot, Pavel Matejka, Mehdi Soufifar, Ricardo Cordoba, Jan Černocký.
Adding Semantics to Clustering Hua Li, Dou Shen, Benyu Zhang, Zheng Chen, Qiang Yang Microsoft Research Asia, Beijing, P.R.China Department of Computer.
Copyright © 2015 by Educational Testing Service. 1 Feature Selection for Automated Speech Scoring Anastassia Loukina, Klaus Zechner, Lei Chen, Michael.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Instance Filtering for Entity Recognition Advisor : Dr.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
1 Web-Page Summarization Using Clickthrough Data* JianTao Sun, Yuchang Lu Dept. of Computer Science TsingHua University Beijing , China Dou Shen,
Efficient Language Model Look-ahead Probabilities Generation Using Lower Order LM Look-ahead Information Langzhou Chen and K. K. Chin Toshiba Research.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
A Baseline System for Speaker Recognition C. Mokbel, H. Greige, R. Zantout, H. Abi Akl A. Ghaoui, J. Chalhoub, R. Bayeh University Of Balamand - ELISA.
Methods for Automatic Evaluation of Sentence Extract Summaries * G.Ravindra +, N.Balakrishnan +, K.R.Ramakrishnan * Supercomputer Education & Research.
Yuya Akita , Tatsuya Kawahara
Extractive Spoken Document Summarization - Models and Features 2007/02/08 Yi-Ting Chen ( 陳怡婷 ) Department of Computer Science & Information Engineering.
Speech Communication Lab, State University of New York at Binghamton Dimensionality Reduction Methods for HMM Phonetic Recognition Hongbing Hu, Stephen.
Voice Activity Detection based on OptimallyWeighted Combination of Multiple Features Yusuke Kida and Tatsuya Kawahara School of Informatics, Kyoto University,
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
1 Latent Concepts and the Number Orthogonal Factors in Latent Semantic Analysis Georges Dupret
A Critique and Improvement of an Evaluation Metric for Text Segmentation A Paper by Lev Pevzner (Harvard University) Marti A. Hearst (UC, Berkeley) Presented.
Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department.
MINIMUM WORD CLASSIFICATION ERROR TRAINING OF HMMS FOR AUTOMATIC SPEECH RECOGNITION Yueng-Tien, Lo Speech Lab, CSIE National.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
Dynamic Tuning Of Language Model Score In Speech Recognition Using A Confidence Measure Sherif Abdou, Michael Scordilis Department of Electrical and Computer.
A Free Software tool for Automatic Tuning of Segmentation Parameters SPT 3.0 Pedro Achanccaray, Victor Ayma, Luis Jimenez, Sergio Garcia, Patrick Happ,
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Phone-Level Pronunciation Scoring and Assessment for Interactive Language Learning Speech Communication, 2000 Authors: S. M. Witt, S. J. Young Presenter:
Correcting Misuse of Verb Forms John Lee, Stephanie Seneff Computer Science and Artificial Intelligence Laboratory, MIT, Cambridge ACL 2008.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Gaussian Mixture Language Models for Speech Recognition Mohamed Afify, Olivier Siohan and Ruhi Sarikaya.
Recent Paper of Md. Akmal Haidar Meeting before ICASSP 2013 報告者:郝柏翰 2013/05/23.
Cross-Dialectal Data Transferring for Gaussian Mixture Model Training in Arabic Speech Recognition Po-Sen Huang Mark Hasegawa-Johnson University of Illinois.
Shadow Detection in Remotely Sensed Images Based on Self-Adaptive Feature Selection Jiahang Liu, Tao Fang, and Deren Li IEEE TRANSACTIONS ON GEOSCIENCE.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
1 Minimum Bayes-risk Methods in Automatic Speech Recognition Vaibhava Geol And William Byrne IBM ; Johns Hopkins University 2003 by CRC Press LLC 2005/4/26.
Arnar Thor Jensson Koji Iwano Sadaoki Furui Tokyo Institute of Technology Development of a Speech Recognition System For Icelandic Using Machine Translated.
Do-Gil Lee1*, Ilhwan Kim1 and Seok Kee Lee2
Efficient Estimation of Word Representation in Vector Space
Multimedia Information Retrieval
Mohamed Kamel Omar and Lidia Mangu ICASSP 2007
A maximum likelihood estimation and training on the fly approach
A Brief Review of Extractive Summarization Research
Presentation transcript:

1 Sentence-extractive automatic speech summarization and evaluation techniques Makoto Hirohata, Yosuke Shinnaka, Koji Iwano, Sadaoki Furui Presented by Yi-Ting Chen

2 Reference Makoto Hirohata, Yosuke Shinnaka, Koji Iwano and Sadaoki Furui, Sentence-extractive automatic speech summarization and evaluation techniques, Speech Communication, In Press, Corrected Proof,, Available online 5 June 2006

3 Outline Introduction Sentence extraction methods Objective evaluation metrics Experiments Conclusion

4 Introduction Why summarize? –Recognition errors cause transcriptions of obtained from spontaneous speech to include irrelevant or incorrect information –Spontaneous speech is ill-formed and usually includes redundant information –Direct transcriptions are therefore not useful They have proposed a two-stage summarization method, but it was confirmed that sentence extraction is more important than sentence compaction This paper investigates and evaluates sentence extraction- based speech summarization techniques under 10% summarization ratio

5 Sentence extraction methods

6 Extraction using sentence location –One hundred and sixty nine presentation were used in the analysis –This result shows that human subjects tend to extract sentence from introduction and conclusion segments under 10% summarization ratio –This is no such tendency at 50%

7 Sentence extraction methods Extraction using sentence location –The introduction and conclusion segments are estimated based on the Hearst method using sentence cohesiveness –The cohesiveness C(r) for each segmentation boundary r is measured by a cosine value: –The segmentation boundary for the end of an introduction part is the beginning of the presentation where the cohesiveness becomes lower than a preset threshold set to 0.1

8 Sentence extraction methods Extraction using confidence and linguistic scores – Confidence score: a logarithmic value of the posterior probability for each transcribed word Linguistic score

9 Sentence extraction methods Extraction using a significance scores Extraction using latent semantic analysis

10 Sentence extraction methods Extraction using dimension reduction based on SVD SVD Dimension reduction Weighted word-frequency vector Weighted singular-value vector Reduced dimension vector

11 Objective evaluation metrics Summarization accuracy –All the human summaries are merged into a single word network –Word accuracy of the automatic summary is then measured as a summarization accuracy in comparison with the closest word string extracted from the word network (SumACCY) –Problem: the variation between manual summaries is so large that the network accepts inappropriate summaries –Using word accuracy obtained by using the manual summaries individually was proposed (SumACCY-E, SumACCY-E/Max, SumACCY-E/avg)

12 Objective evaluation metrics Sentence F-measure –Sentence recall/precision or F-measure is commonly used in evaluating sentence-extraction text summarization. –Since sentence boundaries are not explicitly indicated in input speech –Extraction of a sentence in the recognition result is considered as extraction of one or multiple sentence in the manual summary having an overlap of 50% or more words –F-measure/max, F-measure/ave.

13 Objective evaluation metrics N-gram recall –ROUGE-N is an N-gram recall between an automatic summary and a set of manual summaries. – –S H is a set of manual summaries, S is an individual manual summary, g n is an N-gram, C(g n ) is the number of g n ’s in the manual summary, and C(g n ) is the number of co-occurrences of g n in the manual summary and automatic summary. –1-grams, 2-grams, and 3-grams

14 Experiments Experimental conditions –30 presentations by 20 males and 10 females in the CSJ were automatically summarized at 10% summarization ratio. –Mean word recognition accuracy was 69%. –Sentence boundaries in the recognition results were automatically determined using language models, which achieved 72% recall and 75% precision. Subjective evaluation –180 automatic summaries (30 presentations X 6 summarization methods) were evaluated by 12 human subjects –Methods: SIG, LSA, DIM, SIG+IC, LSA+IC, and DIM+IC –DIM: k set to 5

15 Experiments Correlation between subjective and objective evaluation results

16

17 Experiments Correlation between subjective and objective evaluation results

18

19 Experiments objective evaluation results for various summarizations –The weight factors were optimized a posteriori so that the objective evaluation score got maximized –The parameter K was also set a posteriori to 1 Word recognition accuracy

20 Experiments objective evaluation results for various summarizations Word recognition accuracy

21 Conclusion A sentence extraction method using dimension reduction based on SVD and another method using sentence location information are proposed Among the objective evaluation metrics, summarization accuracy, sentence F-measure, 2 and 3-gram recall were found to the effective under 10% summarization ratio It was confirmed that the summarization method using sentence location improves summarization results