1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.

Slides:



Advertisements
Similar presentations
A Human-Centered Computing Framework to Enable Personalized News Video Recommendation (Oh Jun-hyuk)
Advertisements

Atomatic summarization of voic messages using lexical and prosodic features Koumpis and Renals Presented by Daniel Vassilev.
Linear Model Incorporating Feature Ranking for Chinese Documents Readability Gang Sun, Zhiwei Jiang, Qing Gu and Daoxu Chen State Key Laboratory for Novel.
Chinese Word Segmentation Method for Domain-Special Machine Translation Su Chen; Zhang Yujie; Guo Zhen; Xu Jin’an Beijing Jiaotong University.
Evaluating Search Engine
Speaker: Yun-Nung Chen 陳縕儂 Advisor: Prof. Lin-Shan Lee 李琳山 National Taiwan University Automatic Key Term Extraction and Summarization from Spoken Course.
1 Language Model (LM) LING 570 Fei Xia Week 4: 10/21/2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAA A A.
1 LM Approaches to Filtering Richard Schwartz, BBN LM/IR ARDA 2002 September 11-12, 2002 UMASS.
Introduction to Language Models Evaluation in information retrieval Lecture 4.
1 Today  Tools (Yves)  Efficient Web Browsing on Hand Held Devices (Shrenik)  Web Page Summarization using Click- through Data (Kathy)  On the Summarization.
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
Hierarchical Summaries By: Dawn J. Lawrie University of Massachusetts, Amherst for Search.
DIVINES – Speech Rec. and Intrinsic Variation W.S.May 20, 2006 Richard Rose DIVINES SRIV Workshop The Influence of Word Detection Variability on IR Performance.
Information Retrieval in Practice
Multi-Style Language Model for Web Scale Information Retrieval Kuansan Wang, Xiaolong Li and Jianfeng Gao SIGIR 2010 Min-Hsuan Lai Department of Computer.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
1 Bayesian Learning for Latent Semantic Analysis Jen-Tzung Chien, Meng-Sun Wu and Chia-Sheng Wu Presenter: Hsuan-Sheng Chiu.
1 Wikification CSE 6339 (Section 002) Abhijit Tendulkar.
An Integrated Approach for Arabic-English Named Entity Translation Hany Hassan IBM Cairo Technology Development Center Jeffrey Sorensen IBM T.J. Watson.
AnswerBus Question Answering System Zhiping Zheng School of Information, University of Michigan HLT 2002.
A Survey for Interspeech Xavier Anguera Information Retrieval-based Dynamic TimeWarping.
Summary  The task of extractive speech summarization is to select a set of salient sentences from an original spoken document and concatenate them to.
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
Japanese Spontaneous Spoken Document Retrieval Using NMF-Based Topic Models Xinhui Hu, Hideki Kashioka, Ryosuke Isotani, and Satoshi Nakamura National.
COMPARISON OF A BIGRAM PLSA AND A NOVEL CONTEXT-BASED PLSA LANGUAGE MODEL FOR SPEECH RECOGNITION Md. Akmal Haidar and Douglas O’Shaughnessy INRS-EMT,
Yun-Nung (Vivian) Chen, Yu Huang, Sheng-Yi Kong, Lin-Shan Lee National Taiwan University, Taiwan.
1 Sentence-extractive automatic speech summarization and evaluation techniques Makoto Hirohata, Yosuke Shinnaka, Koji Iwano, Sadaoki Furui Presented by.
DISCRIMINATIVE TRAINING OF LANGUAGE MODELS FOR SPEECH RECOGNITION Hong-Kwang Jeff Kuo, Eric Fosler-Lussier, Hui Jiang, Chin-Hui Lee ICASSP 2002 Min-Hsuan.
Web-Assisted Annotation, Semantic Indexing and Search of Television and Radio News (proceedings page 255) Mike Dowman Valentin Tablan Hamish Cunningham.
Automatic Speech Recognition: Conditional Random Fields for ASR Jeremy Morris Eric Fosler-Lussier Ray Slyh 9/19/2008.
Binxing Jiao et. al (SIGIR ’10) Presenter : Lin, Yi-Jhen Advisor: Dr. Koh. Jia-ling Date: 2011/4/25 VISUAL SUMMARIZATION OF WEB PAGES.
8.0 Search Algorithms for Speech Recognition References: of Huang, or of Becchetti, or , of Jelinek 4. “ Progress.
1 Boostrapping language models for dialogue systems Karl Weilhammer, Matthew N Stuttle, Steve Young Presenter: Hsuan-Sheng Chiu.
Collocations and Information Management Applications Gregor Erbach Saarland University Saarbrücken.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
A Scalable Machine Learning Approach for Semi-Structured Named Entity Recognition Utku Irmak(Yahoo! Labs) Reiner Kraft(Yahoo! Inc.) WWW 2010(Information.
Cluster-specific Named Entity Transliteration Fei Huang HLT/EMNLP 2005.
Chapter 23: Probabilistic Language Models April 13, 2004.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Alignment of Bilingual Named Entities in Parallel Corpora Using Statistical Model Chun-Jen Lee Jason S. Chang Thomas C. Chuang AMTA 2004.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
National Taiwan University, Taiwan
Improving Named Entity Translation Combining Phonetic and Semantic Similarities Fei Huang, Stephan Vogel, Alex Waibel Language Technologies Institute School.
A Word Clustering Approach for Language Model-based Sentence Retrieval in Question Answering Systems Saeedeh Momtazi, Dietrich Klakow University of Saarland,Germany.
Creating Subjective and Objective Sentence Classifier from Unannotated Texts Janyce Wiebe and Ellen Riloff Department of Computer Science University of.
Interactive Navigation across Spoken Document Archives with Topic Hierarchies Constructed by Properly Ranked Key Terms Speaker: Yi-Cheng Pan (Thomas) Advisor:
Latent Topic Modeling of Word Vicinity Information for Speech Recognition Kuan-Yu Chen, Hsuan-Sheng Chiu, Berlin Chen ICASSP 2010 Hao-Chin Chang Department.
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Relevance Language Modeling For Speech Recognition Kuan-Yu Chen and Berlin Chen National Taiwan Normal University, Taipei, Taiwan ICASSP /1/17.
A New Approach for English- Chinese Named Entity Alignment Donghui Feng Yayuan Lv Ming Zhou USC MSR Asia EMNLP-04.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Copyright © 2013 by Educational Testing Service. All rights reserved. Evaluating Unsupervised Language Model Adaption Methods for Speaking Assessment ShaSha.
Finding document topics for improving topic segmentation Source: ACL2007 Authors: Olivier Ferret (18 route du Panorama, BP6) Reporter:Yong-Xiang Chen.
10.0 Latent Semantic Analysis for Linguistic Processing References : 1. “Exploiting Latent Semantic Information in Statistical Language Modeling”, Proceedings.
Divided Pretreatment to Targets and Intentions for Query Recommendation Reporter: Yangyang Kang /23.
A Generation Model to Unify Topic Relevance and Lexicon-based Sentiment for Opinion Retrieval Min Zhang, Xinyao Ye Tsinghua University SIGIR
Pastra and Saggion, EACL 2003 Colouring Summaries BLEU Katerina Pastra and Horacio Saggion Department of Computer Science, Natural Language Processing.
Single Document Key phrase Extraction Using Neighborhood Knowledge.
Confidence Measures As a Search Guide In Speech Recognition Sherif Abdou, Michael Scordilis Department of Electrical and Computer Engineering, University.
A Maximum Entropy Language Model Integrating N-grams and Topic Dependencies for Conversational Speech Recognition Sanjeev Khudanpur and Jun Wu Johns Hopkins.
Maximum Entropy techniques for exploiting syntactic, semantic and collocational dependencies in Language Modeling Sanjeev Khudanpur, Jun Wu Center for.
Author :K. Thambiratnam and S. Sridharan DYNAMIC MATCH PHONE-LATTICE SEARCHES FOR VERY FAST AND ACCURATE UNRESTRICTED VOCABULARY KEYWORD SPOTTING Reporter.
Recent Paper of Md. Akmal Haidar Meeting before ICASSP 2013 報告者:郝柏翰 2013/05/23.
Linguistic knowledge for Speech recognition
Speaker : chia hua Authors : Long Qin, Ming Sun, Alexander Rudnicky
Multimedia Information Retrieval
Mohamed Kamel Omar and Lidia Mangu ICASSP 2007
Content Augmentation for Mixed-Mode News Broadcasts Mike Dowman
Anthor: Andreas Tsiartas, Prasanta Kumar Ghosh,
Presentation transcript:

1 ICASSP Paper Survey Presenter: Chen Yi-Ting

2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis (PLSA) Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA) Topic and Stylistic Adaptation for Speech Summarisation Automatic Sentence Segmentation of Speech for Automatic Summarization

3 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis (PLSA)

4 In this paper, using a “dynamic key term lexicon” automatically extracted from the ever-changing document archives as an extra feature set in the retrieval task.

5 An important part of the proposed approach is the automatic key term extraction from the archives The second important part of the proposed approach is the key term recognition from the user spoken query –Here special approaches to recognize correctly the key term in the user query were developed, including emphasizing the possible key term candidates during search through the phone lattice, and key term matching using a phone similarity matrix including two different distance measures. –Two different versions of models can be used: the general lexicon including all terms except those stop terms deleted and the other based on the much smaller but semantically rich key term lexicon

6 Named Entity Recogniion –The first is to recognize the NEs from a text document (or the transcription of a spoken document) using global information –The second special approach used here is for spoken documents to recover the OOV NEs using external knowledge Key Term extraction by term entropy based on PLSA

7 Key term recognition from user query –The user spoken query is transcribed not only into a word graph as usual recognition process, but into a phone lattice as well –Then matching the phone lattice with the phone sequences of the key terms in the dynamic lexicon using dynamic programming (threshold) –The price paid here is of course the overall word error rate may be increased The experimental conditions –Word error rates 27% 、 character error rates 14.29% 、 syllable error rate 8.91% –32 topics were used in PLSA modeling –1000new stories (test set) 50 queries –The length of the queries is roughly 8-11 words –A lexicon of word was used here –A total of 1708 NE were obtained (from 9836 news stories) –Picked up the top 2000 terms ranked by term entropy

8 Experimental Results

9 Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA)

10 where some statistical measure s(tj) (such as TF/IDF or the like) linguistic measure l(tj) (e.g., named entities) c(tj) is calculated from the confidence score g(tj) is N-gram score for the term tj b(S) is calculated from the grammatical structure of the sentence S λ1, λ2, λ3, λ4 and λ5 are weighting parameters Two useful measures, referred to as topic significance and term entropy in this paper are proposed based on the PLSA modeling to determine the terms and thus sentences important for the document which can then be used to construct the summary The statistical measure s(tj) which has been proved extremely useful is called “significance score”:

11 Topic significance –The topic significance score of a term tj with respect to a topic Tk –The statistical measure: Term Entropy – is a scaling factor

12 Experiments configuration –The test corpus included 200 news stories broadcast –Word accuracy 66.46% 、 character accuracy 74.95% 、 syllable accuracy 81.70% –Sentence recall/precision is the evaluation metric for automatic summarization of documents

13 Experiments configuration

14 Topic and Stylistic Adaptation for Speech Summarisation

15 In this paper they investigate LiM topic and stylistic adaptation using combinations of LiMs each trained on different adaptation data Focusing on adapting the linguistic component, which is not related at all to the language model used during the recognition process, to make it more suited for the summarisation task Experiments were performed both on spontaneous speech, using 9 talks from the Translanguage English Database (TED) corpus, and speech read from text, using 5 talks from CNN broadcast news from 1995 The measure of summary quality used in this paper is summarisation accuracy (SumACCY)

16 Automatic speech summarisation system

17 Summarisation Method Important sentences are first extracted according to the following score for each sentence, obtained from the automatic speech recognition out (ASR) Starting with a baseline LiM (LiM B ) we perform LiM adaptation by linearly interpolating the baseline model with other component models trained on different data Where Different types of component LiM are built, coming from different sources of data, and using either unigram, bigram or trigram information

18 Experimental Setup –Due to lack of data we had to use the talks both for development and evaluation with a rotating form of cross-validation: all talks but one are used for development, the remaining talk being used for testing –Summaries from the development talks are generated automatically by the system using different sets of parameters –For the TED data, two type of component linguistic models: The first type are built on the small corpus of hand-made summaries, made for the same summarisation ratio. The second type are built from the papers in the conference proceeding for talk we want to summarise –For the CNN data, one type of component linguistic models the small corpus of hand-made summaries

19 Experimental Setup –Reference results: random summarisation, the humman summaries and the baseline CNNTED

20

21 Automatic Sentence Segmentation of Speech for Automatic Summarization

22 This paper presents an automatic sentence segmentation method for an automatic speech summarization system The segmentation method is based on combining word- and class-based statistical language models to predict sentence and non-sentence boundaries Studying both the effect of the segmentation on the sentence segmentation system itself and the effect of the segmentation on the summarization accuracy To judge the quality of the sentence segmentation we used the F-measure metrics

23 Automatic sentence segmentation This probability was combined with the matching recursive path probability from Three LMs were used in sentence segmentation, two word- based LMs and a class-based LM The LMs were combined by linear interpolation as follows:

24 Experimental results