Pruning Analysis for the Position Specific Posterior Lattices for Spoken Document Search Jorge Silva University of Southern California Ciprian Chelba and.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Information Retrieval in Practice
Metadata in Carrot II Current metadata –TF.IDF for both documents and collections –Full-text index –Metadata are transferred between different nodes Potential.
Chapter 5: Introduction to Information Retrieval
Multimedia Database Systems
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Representing and Querying Correlated Tuples in Probabilistic Databases
File Processing : Hash 2015, Spring Pusan National University Ki-Joune Li.
Lattices Segmentation and Minimum Bayes Risk Discriminative Training for Large Vocabulary Continuous Speech Recognition Vlasios Doumpiotis, William Byrne.
Bag-Of-Word normalized n-gram models ISCA 2008 Abhinav Sethy, Bhuvana Ramabhadran IBM T. J. Watson Research Center Yorktown Heights, NY Presented by Patty.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
Evaluating Search Engine
DYNAMIC ELEMENT RETRIEVAL IN A STRUCTURED ENVIRONMENT MAYURI UMRANIKAR.
MANISHA VERMA, VASUDEVA VARMA PATENT SEARCH USING IPC CLASSIFICATION VECTORS.
Spoken Document Retrieval and Browsing Ciprian Chelba.
Presented by Zeehasham Rasheed
Scalable Text Mining with Sparse Generative Models
Xiaomeng Su & Jon Atle Gulla Dept. of Computer and Information Science Norwegian University of Science and Technology Trondheim Norway June 2004 Semantic.
1 The Mystery of Cooperative Web Caching 2 b b Web caching : is a process implemented by a caching proxy to improve the efficiency of the web. It reduces.
Aparna Kulkarni Nachal Ramasamy Rashmi Havaldar N-grams to Process Hindi Queries.
DIVINES – Speech Rec. and Intrinsic Variation W.S.May 20, 2006 Richard Rose DIVINES SRIV Workshop The Influence of Word Detection Variability on IR Performance.
IBM Haifa Research Lab © 2008 IBM Corporation Retrieving Spoken Information by Combining Multiple Speech Transcription Methods Jonathan Mamou Joint work.
Multi-Style Language Model for Web Scale Information Retrieval Kuansan Wang, Xiaolong Li and Jianfeng Gao SIGIR 2010 Min-Hsuan Lai Department of Computer.
Lightly Supervised and Unsupervised Acoustic Model Training Lori Lamel, Jean-Luc Gauvain and Gilles Adda Spoken Language Processing Group, LIMSI, France.
Probabilistic Model for Definitional Question Answering Kyoung-Soo Han, Young-In Song, and Hae-Chang Rim Korea University SIGIR 2006.
Automatic Spoken Document Processing for Retrieval and Browsing Zahra Ahmadi.
Improved search for Socially Annotated Data Authors: Nikos Sarkas, Gautam Das, Nick Koudas Presented by: Amanda Cohen Mostafavi.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
1 Efficient Search Ranking in Social Network ACM CIKM2007 Monique V. Vieira, Bruno M. Fonseca, Rodrigo Damazio, Paulo B. Golgher, Davi de Castro Reis,
Bayesian Extension to the Language Model for Ad Hoc Information Retrieval Hugo Zaragoza, Djoerd Hiemstra, Michael Tipping Presented by Chen Yi-Ting.
Japanese Spontaneous Spoken Document Retrieval Using NMF-Based Topic Models Xinhui Hu, Hideki Kashioka, Ryosuke Isotani, and Satoshi Nakamura National.
Summarization of XML Documents K Sarath Kumar. Outline I.Motivation II.System for XML Summarization III.Ranking Model and Summary Generation IV.Example.
Ranking in Information Retrieval Systems Prepared by: Mariam John CSE /23/2006.
HIERARCHICAL SEARCH FOR LARGE VOCABULARY CONVERSATIONAL SPEECH RECOGNITION Author :Neeraj Deshmukh, Aravind Ganapathiraju and Joseph Picone.
Efficient Instant-Fuzzy Search with Proximity Ranking Authors: Inci Centidil, Jamshid Esmaelnezhad, Taewoo Kim, and Chen Li IDCE Conference 2014 Presented.
The Anatomy of a Large-Scale Hyper textual Web Search Engine S. Brin, L. Page Presenter :- Abhishek Taneja.
Handing Uncertain Observations in Unsupervised Topic-Mixture Language Model Adaptation Ekapol Chuangsuwanich 1, Shinji Watanabe 2, Takaaki Hori 2, Tomoharu.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
1 Sentence Extraction-based Presentation Summarization Techniques and Evaluation Metrics Makoto Hirohata, Yousuke Shinnaka, Koji Iwano and Sadaoki Furui.
Personalization with user’s local data Personalizing Search via Automated Analysis of Interests and Activities 1 Sungjick Lee Department of Electrical.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Chapter 5 Ranking with Indexes 1. 2 More Indexing Techniques n Indexing techniques:  Inverted files - best choice for most applications  Suffix trees.
Web Information Retrieval Prof. Alessandro Agostini 1 Context in Web Search Steve Lawrence Speaker: Antonella Delmestri IEEE Data Engineering Bulletin.
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
A DYNAMIC APPROACH TO THE SELECTION OF HIGH ORDER N-GRAMS IN PHONOTACTIC LANGUAGE RECOGNITION Mikel Penagarikano, Amparo Varona, Luis Javier Rodriguez-
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
Date: 2012/08/21 Source: Zhong Zeng, Zhifeng Bao, Tok Wang Ling, Mong Li Lee (KEYS’12) Speaker: Er-Gang Liu Advisor: Dr. Jia-ling Koh 1.
Relevance Language Modeling For Speech Recognition Kuan-Yu Chen and Berlin Chen National Taiwan Normal University, Taipei, Taiwan ICASSP /1/17.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Copyright © 2013 by Educational Testing Service. All rights reserved. Evaluating Unsupervised Language Model Adaption Methods for Speaking Assessment ShaSha.
Ranking of Database Query Results Nitesh Maan, Arujn Saraswat, Nishant Kapoor.
GENERATING RELEVANT AND DIVERSE QUERY PHRASE SUGGESTIONS USING TOPICAL N-GRAMS ELENA HIRST.
1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis.
Author :K. Thambiratnam and S. Sridharan DYNAMIC MATCH PHONE-LATTICE SEARCHES FOR VERY FAST AND ACCURATE UNRESTRICTED VOCABULARY KEYWORD SPOTTING Reporter.
Discriminative n-gram language modeling Brian Roark, Murat Saraclar, Michael Collins Presented by Patty Liu.
Recent Paper of Md. Akmal Haidar Meeting before ICASSP 2013 報告者:郝柏翰 2013/05/23.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Bayes Rule Mutual Information Conditional.
XRANK: RANKED KEYWORD SEARCH OVER XML DOCUMENTS Lin Guo Feng Shao Chavdar Botev Jayavel Shanmugasundaram Abhishek Chennaka, Alekhya Gade Advanced Database.
1 Minimum Bayes-risk Methods in Automatic Speech Recognition Vaibhava Geol And William Byrne IBM ; Johns Hopkins University 2003 by CRC Press LLC 2005/4/26.
Collection Fusion in Carrot2
2 Research Department, iFLYTEK Co. LTD.
SEARCH ENGINES & WEB CRAWLER Akshay Ghadge Roll No: 107.
Compact Query Term Selection Using Topically Related Text
2018, Spring Pusan National University Ki-Joune Li
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Presenter : Jen-Wei Kuo
Information Retrieval and Web Design
Presentation transcript:

Pruning Analysis for the Position Specific Posterior Lattices for Spoken Document Search Jorge Silva University of Southern California Ciprian Chelba and Alex Acero Microsoft Corporation Presented by Shih-Hung Liu 2006/06/12 ICASSP2006

2 Outline Introduction Position specific posterior lattices (PSPL) Spoken document indexing and search using PSPL Pruning techniques –Relative pruning –Absolute pruning Experiments Conclusions

3 Introduction Speech search has not received much attention due to the fact that large collections of un-transcribed spoken material have not available, mostly due to storage constraints As storage becomes cheaper, the availability of large collections of spoken documents is limited strictly by the lack of adequate technology to exploit it them Manually transcribing speech is expensive and sometimes outright impossible due to privacy concerns This leads us to exploring an automatic approach for searching and navigating spoken document collections

4 Introduction The main research aiming at spoken document retrieval (SDR) was centered around the SDR-TREC evaluation –SDR systems index ASR 1-best output and use mean average precision (MAP) However there are shortcomings to the SDR-TREC framework –the recognizers were heavily tuned for the domain leading to very good ASR performance The effect of higher WER needs to be explored

5 Position specific posterior lattices The early Google approach shows that context and proximity issues were to be important factors for doing the relevance ranking score If we want to extend this direction to spoken documents, we are faced with a dilemma –using 1-best ASR output is suboptimal due to high WER –ASR lattices do have much better WER, but the position information needed for rescoring a given word hit is not readily available

6 Position specific posterior lattices Position specific posterior Lattice (PSPL) was proposed as a way to extend the key-word search paradigm from text documents to spoken documents in scenarios with high WER The approach calculates posterior probabilities of words at a given integer position – soft indexing – as a way to model the uncertainty of spoken documents

7 Position specific posterior lattices Since it is possible that more than one path contains the same word in the same position, one would need to sum over all possible paths in a lattice that contain a given word at given position The computation for the backward pass stays unchanged, whereas during the forward pass one needs to split the forward probability arriving at a given node,, according to the length of the partial paths

8 Position specific posterior lattices Using the forward-backward variables the posterior probability of a given word occurring at a given position in the lattices can be easily calculated : Finally, the Position Specific Posterior Lattice (PSPL) is a representation of the distribution : for each bin store the words along with their posterior probability

9 Spoken document indexing and search using PSPL In our case the speech content of a typical spoken document was approximately 1hr long; it is customary to segment a given speech file in shorter segments A spoken document thus consists of an ordered list of segments. For each segment we generate a corresponding PSPL lattices

10 Spoken document indexing and search using PSPL For all query terms, a 1-gram score is calculated by summing the PSPL posterior probability across all segment and position The logarithmic tapering off is used for discounting the effect of large counts in a given document

11 Spoken document indexing and search using PSPL We calculate an expected tapered-count for each N-gram in the query The different proximity types :

12 Pruning techniques The idea is to explore how the PSPL ranking performance behaves as a function of the level uncertainty transferred from the ASR part For doing that, the idea is to prune the PSPL probability distribution before the generation of the N-gram expected counts

13 Relative pruning For a given PSPL position bin, the relative pruning first finds the most likely entry given by The remaining entries are renormalized to make it a proper probability mass function and used to calculate the expected N-gram counts

14 Absolute pruning In this case, the PSPL framework considers the PSPL entries whose log-probability is higher than an absolute threshold The threshold is absolute and consequently when is relatively close to zero, the PSPL contains only the bin entries that have high level of confidence, and some position bins become empty

15 Experiments Corpus –iCampus corpus prepared by MIT CSAIL –169 hrs of lecture material recorded in the classroom 90 Lectures (78.5 hrs) 79 Assorted MIT World seminars (89.9 hrs) –The speech style is in between planned and spontaneous

16 Experiments

17 Experiments

18 Conclusions The PSPL framework provides an alternative way of dealing with the document’s content uncertainty for spoken document retrieval The proposed pruning techniques provide ways of controlling the number of possible entries in the PSPL, based on their log-probability