Estimate the Number of Relevant Images Using Two-Order Markov Chain Presented by: WANG Xiaoling Supervisor: Clement LEUNG.

Slides:



Advertisements
Similar presentations
Stationary Probability Vector of a Higher-order Markov Chain By Zhang Shixiao Supervisors: Prof. Chi-Kwong Li and Dr. Jor-Ting Chan.
Advertisements

Aggregating local image descriptors into compact codes
Worksheet I. Exercise Solutions Ata Kaban School of Computer Science University of Birmingham.
Matrices, Digraphs, Markov Chains & Their Use by Google Leslie Hogben Iowa State University and American Institute of Mathematics Leslie Hogben Iowa State.
CIVL3310 STRUCTURAL ANALYSIS Professor CC Chang Introduction to Matrix based techniques.
The Rate of Concentration of the stationary distribution of a Markov Chain on the Homogenous Populations. Boris Mitavskiy and Jonathan Rowe School of Computer.
Matrices, Digraphs, Markov Chains & Their Use. Introduction to Matrices  A matrix is a rectangular array of numbers  Matrices are used to solve systems.
Estimation of the Number of Relevant Images in Infinite Databases Presented by: Xiaoling Wang Supervisor: Prof. Clement Leung.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Hinrich Schütze and Christina Lioma Lecture 12: Language Models for IR
Evaluating Evaluation Measure Stability Authors: Chris Buckley, Ellen M. Voorhees Presenters: Burcu Dal, Esra Akbaş.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
CSE 3504: Probabilistic Analysis of Computer Systems Topics covered: Continuous time Markov chains (Sec )
Kurtis Cahill James Badal.  Introduction  Model a Maze as a Markov Chain  Assumptions  First Approach and Example  Second Approach and Example 
Multimedia Databases SVD II. Optimality of SVD Def: The Frobenius norm of a n x m matrix M is (reminder) The rank of a matrix M is the number of independent.
Expectation Maximization Method Effective Image Retrieval Based on Hidden Concept Discovery in Image Database By Sanket Korgaonkar Masters Computer Science.
Face Recognition Using Embedded Hidden Markov Model.
. Class 5: HMMs and Profile HMMs. Review of HMM u Hidden Markov Models l Probabilistic models of sequences u Consist of two parts: l Hidden states These.
Multimedia Databases SVD II. SVD - Detailed outline Motivation Definition - properties Interpretation Complexity Case studies SVD properties More case.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference (Sec. )
Link Analysis, PageRank and Search Engines on the Web
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference.
Student: Hsu-Yung Cheng Advisor: Jenq-Neng Hwang, Professor
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Discrete time Markov chains (Sec )
Affinity Rank Yi Liu, Benyu Zhang, Zheng Chen MSRA.
Probabilistic Model of Sequences Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán)
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Discrete time Markov chains (Sec. 7.1)
Division in the Integers Theorem 1. If n and m are integers and n>0, then m=qn+r for unique integers q and r with 0
Motivation When searching for information on the WWW, user perform a query to a search engine. The engine return, as the query’s result, a list of Web.
Approximate Inference 2: Monte Carlo Markov Chain
The effect of New Links on Google Pagerank By Hui Xie Apr, 07.
Properties of Real Numbers
1 Data Mining over the Deep Web Tantan Liu, Gagan Agrawal Ohio State University April 12, 2011.
Performance of Recommender Algorithms on Top-N Recommendation Tasks RecSys 2010 Intelligent Database Systems Lab. School of Computer Science & Engineering.
1 Location Estimation in ZigBee Network Based on Fingerprinting Department of Computer Science and Information Engineering National Cheng Kung University,
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
PERSONALIZED SEARCH Ram Nithin Baalay. Personalized Search? Search Engine: A Vital Need Next level of Intelligent Information Retrieval. Retrieval of.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
Methods Model. The TECOS model is used as forward model to simulate carbon transfer among the carbon pools (Fig.1). In the model, ecosystem is simplified.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
Optimization of a Finite State Space for Information Retrieval Qiang Huang School of Computing, Robert Gordon University.
Search engines are used to for looking for documents. They compile their databases by employing "spiders" or "robots" to crawl through web space from.
Detecting Dominant Locations from Search Queries Lee Wang, Chuang Wang, Xing Xie, Josh Forman, Yansheng Lu, Wei-Ying Ma, Ying Li SIGIR 2005.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Properties and Mental Computation p. 80. Math talk What are some math properties that we use? Why do you think we have them? Do you ever use them?
Chapter 13 Multiple Regression
1 Selectivity Estimation for Exclusive Query Translation in Deep Web Data Integration Fangjiao Jiang Renmin University of China Joint work with Weiyi Meng.
1 Language Specific Crawler for Myanmar Web Pages Pann Yu Mon Management and Information System Engineering Department Nagaoka University of Technology,
Adjustment of Level Nets. Introduction In this chapter we will deal with differential leveling only In SUR2101 you learned how to close and adjust a level.
Dongfang Xu School of Information
PERSONALIZED DIVERSIFICATION OF SEARCH RESULTS Date: 2013/04/15 Author: David Vallet, Pablo Castells Source: SIGIR’12 Advisor: Dr.Jia-ling, Koh Speaker:
By: Jesse Ehlert Dustin Wells Li Zhang Iterative Aggregation/Disaggregation(IAD)
ST3236: Stochastic Process Tutorial 6
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
黃福銘 (Angus F.M. Huang) ANTS Lab, IIS, Academia Sinica Exploring Spatial-Temporal Trajectory Model for Location.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Introduction. We want to see if there is any relationship between the results on exams and the amount of hours used for studies. Person ABCDEFGHIJ Hours/
Behavior Recognition Based on Machine Learning Algorithms for a Wireless Canine Machine Interface Students: Avichay Ben Naim Lucie Levy 14 May, 2014 Ort.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
Search Engines and Link Analysis on the Web
MATHEMATICS Matrix Multiplication
Matrix Multiplication
A research literature search engine with abbreviation recognition
Chapter 9: One- and Two-Sample Estimation Problems:
Discrete-time markov chain (continuation)
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Properties of Real Numbers
Properties of Numbers Review Problems.
Presentation transcript:

Estimate the Number of Relevant Images Using Two-Order Markov Chain Presented by: WANG Xiaoling Supervisor: Clement LEUNG

Outline Introduction Introduction Objective Objective Methodology Methodology Experiment Results Experiment Results Conclusion and Future Work Conclusion and Future Work

Introduction Large collections of images have been made available on web. Large collections of images have been made available on web. Retrieval effectiveness becomes one of the most important parameters to measure the performance of image retrieval systems. Retrieval effectiveness becomes one of the most important parameters to measure the performance of image retrieval systems.

Measures: Measures: Precision Precision Recall Recall Significant Challenge: the total number of relevant images is not directly observable Significant Challenge: the total number of relevant images is not directly observable

Basic Models Basic Models Regression Model Regression Model Markov Chain Markov Chain Two-Order Markov Chain Two-Order Markov Chain

Objective To investigate the probabilistic behavior of the distribution of relevant images among the returned results for the image search engines using two-order markov chain To investigate the probabilistic behavior of the distribution of relevant images among the returned results for the image search engines using two-order markov chain

Methodology Test Image Search Engine: Test Image Search Engine: Query Design Query Design 70% provided by authors 70% provided by authors One word query One word query Two word query Two word query Three word query Three word query 30% suggestive term 30% suggestive term Suggestive term with largest returned results Suggestive term with largest returned results Suggestive term with least returned results Suggestive term with least returned results

Methodology Database Setup: Database Setup: Stochastic process {X 1, X 2, …, X J } Stochastic process {X 1, X 2, …, X J } where X J denotes the aggregate relevance of all the images in page J where X J denotes the aggregate relevance of all the images in page J Equation: Equation: where Y Ji =1 if the i th image on page J is relevant, and Y Ji =0 if the i th image on page J is not relevant. where Y Ji =1 if the i th image on page J is relevant, and Y Ji =0 if the i th image on page J is not relevant.

Page J XJXJXJXJ

Forecast Using Two-Order Markov Chain Forecast Using Two-Order Markov Chain Markov Chain: Stochastic process {X J, J≥1} with state space S={0,1,2,…20} , Markov Chain: Stochastic process {X J, J≥1} with state space S={0,1,2,…20} , Two-Order Markov Chain: State space change to S 2 , Two-Order Markov Chain: State space change to S 2 , Forecast the state probability distribution of next page π(J) based on the original state probability distribution π(1) and transition probability matrix P. An Example Forecast the state probability distribution of next page π(J) based on the original state probability distribution π(1) and transition probability matrix P. An ExampleAn ExampleAn Example Model Test Model Test Mean Absolute Error Mean Absolute Error

Experiment Results Forecast Results Using Two-Order Markov Chain Forecast Results Using Two-Order Markov Chain PageGoogleYahooBing

Test Results--Google Test Results--Google

Test Results--Yahoo Test Results--Yahoo

Test Results--Bing Test Results--Bing

Measure for Forecast Accuracy Measure for Forecast Accuracy Mean Absolute Deviation (MAD): Mean Absolute Deviation (MAD):One-wordTwo-wordThree-wordGoogle Yahoo Bing

Comparative Results Comparative Results Best Model: Two- Order Markov Chain Best Model: Two- Order Markov Chain Worst Model: Regression Model Worst Model: Regression Model

Conclusion Two-Order Markov Chain could well represent the distribution of relevant images among the results pages for the major web image search engine. Two-Order Markov Chain could well represent the distribution of relevant images among the results pages for the major web image search engine. Two-Order Markov Chain is the best model among three models we have worked. Two-Order Markov Chain is the best model among three models we have worked.

Future Work Our future work will try to apply Hidden Markov Chain to this topic Our future work will try to apply Hidden Markov Chain to this topic

Thank you! Q & A

Two-Order Markov Chain An example (cont ’ ) Suppose the stochastic process {X t, t>=0} with a state space S={A, B, C} Suppose the stochastic process {X t, t>=0} with a state space S={A, B, C} As to two-order Markov chain, the state space: As to two-order Markov chain, the state space: S 2 ={AA, AB, AC, BA, BB, BC, CA, CB, CC} S 2 ={AA, AB, AC, BA, BB, BC, CA, CB, CC} The state probabilities distribution of period zero: The state probabilities distribution of period zero:  (0)= (  AA,  AB,  AC,  BA,  BB,  BC,  CA,  CB,  CC )  (0)= (  AA,  AB,  AC,  BA,  BB,  BC,  CA,  CB,  CC )

An example (cont ’ ) The transition probability matrix: The transition probability matrix: P AA,BA =0

An example Therefore, the probability distribution of states for page J will be compute as: Therefore, the probability distribution of states for page J will be compute as: π(J)=π(J-1)*P π(J)=π(J-1)*P [Return] [Return]Return