Millions of Databases: Which are Trustworthy and Relevant?

Slides:



Advertisements
Similar presentations
Accurately Interpreting Clickthrough Data as Implicit Feedback Joachims, Granka, Pan, Hembrooke, Gay Paper Presentation: Vinay Goel 10/27/05.
Advertisements

Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Lindsey Bleimes Charlie Garrod Adam Meyerson
Risk Modeling The Tropos Approach PhD Lunch Meeting 07/07/2005 Yudistira Asnar –
Web Information Retrieval
Ranking Tweets Considering Trust and Relevance Srijith Ravikumar,Raju Balakrishnan, and Subbarao Kambhampati Arizona State University 1.
Search Advertising These slides are modified from those by Anand Rajaram & Jeff UllmanAnand Rajaram & Jeff Ullman.
Dr. Subbarao Kambhampati
Dong Liu Xian-Sheng Hua Linjun Yang Meng Weng Hong-Jian Zhang.
1 The PageRank Citation Ranking: Bring Order to the web Lawrence Page, Sergey Brin, Rajeev Motwani and Terry Winograd Presented by Fei Li.
Case Study: BibFinder BibFinder: A popular CS bibliographic mediator –Integrating 8 online sources: DBLP, ACM DL, ACM Guide, IEEE Xplore, ScienceDirect,
Trust and Profit Sensitive Ranking for Web Databases and On-line Advertisements Raju Balakrishnan (Arizona State University)
Optimal Ad Ranking for Profit Maximization Raju Balakrishnan (Arizona State University) Subbarao Kambhampati (Arizona State University) TexPoint fonts.
Trust and Profit Sensitive Ranking for Web Databases and On-line Advertisements Raju Balakrishnan (PhD Dissertation Defense) Committee: Subbarao.
Trust and Profit Sensitive Ranking for Web Databases and On-line Advertisements Raju Balakrishnan (PhD Proposal Defense) Committee: Subbarao Kambhampati.
Evaluating Search Engine
Topic-Sensitive SourceRank: Agreement Based Source Selection for the Multi-Topic Deep Web Integration Manishkumar Jha Raju Balakrishnan Subbarao Kambhampati.
Maggie Zhou COMP 790 Data Mining Seminar, Spring 2011
Supporting Queries with Imprecise Constraints Ullas Nambiar Dept. of Computer Science University of California, Davis Subbarao Kambhampati Dept. of Computer.
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
CS246: Page Selection. Junghoo "John" Cho (UCLA Computer Science) 2 Page Selection Infinite # of pages on the Web – E.g., infinite pages from a calendar.
1 Information Management on the World-Wide Web Junghoo “John” Cho UCLA Computer Science.
Zdravko Markov and Daniel T. Larose, Data Mining the Web: Uncovering Patterns in Web Content, Structure, and Usage, Wiley, Slides for Chapter 1:
Sigir’99 Inside Internet Search Engines: Search Jan Pedersen and William Chang.
Topic-Sensitive PageRank Taher H. Haveliwala. PageRank Importance is propagated A global ranking vector is pre-computed.
(hyperlink-induced topic search)
1 COMP4332 Web Data Thanks for Raymond Wong’s slides.
Web Search – Summer Term 2006 II. Information Retrieval (Basics Cont.) (c) Wolfgang Hürst, Albert-Ludwigs-University.
Λ14 Διαδικτυακά Κοινωνικά Δίκτυα και Μέσα
Graph 8 a. Graph b. Domain _________ c. Range __________
Trust and Profit Sensitive Ranking for Web Databases and On-line Advertisements Raju Balakrishnan (Arizona State University)
1 Scopus as a Research Tool March Why Scopus?  A comprehensive abstract and citation database of peer-reviewed literature and quality web sources.
Group Recommendations with Rank Aggregation and Collaborative Filtering Linas Baltrunas, Tadas Makcinskas, Francesco Ricci Free University of Bozen-Bolzano.
1 Applying Collaborative Filtering Techniques to Movie Search for Better Ranking and Browsing Seung-Taek Park and David M. Pennock (ACM SIGKDD 2007)
PERSONALIZED SEARCH Ram Nithin Baalay. Personalized Search? Search Engine: A Vital Need Next level of Intelligent Information Retrieval. Retrieval of.
SourceRank: Relevance and Trust Assessment for Deep Web Sources Based on Inter-Source Agreement Raju Balakrishnan, Subbarao Kambhampati Arizona State University.
Fall 2006 Davison/LinCSE 197/BIS 197: Search Engine Strategies 2-1 How Search Engines Work Today we show how a search engine works  What happens when.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Glasgow 02/02/04 NN k networks for content-based image retrieval Daniel Heesch.
SWE.org. Not what we wanted to hear … but we needed to listen Some choice findings from our initial member survey “Make the material more user friendly.
The PageRank Citation Ranking: Bringing Order to the Web Lawrence Page, Sergey Brin, Rajeev Motwani, Terry Winograd Presented by Anca Leuca, Antonis Makropoulos.
Center for E-Business Technology Seoul National University Seoul, Korea BrowseRank: letting the web users vote for page importance Yuting Liu, Bin Gao,
2010 © University of Michigan 1 DivRank: Interplay of Prestige and Diversity in Information Networks Qiaozhu Mei 1,2, Jian Guo 3, Dragomir Radev 1,2 1.
Distributed Information Retrieval Server Ranking for Distributed Text Retrieval Systems on the Internet B. Yuwono and D. Lee Siemens TREC-4 Report: Further.
Improving Web Search Results Using Affinity Graph Benyu Zhang, Hua Li, Yi Liu, Lei Ji, Wensi Xi, Weiguo Fan, Zheng Chen, Wei-Ying Ma Microsoft Research.
Evaluation of Recommender Systems Joonseok Lee Georgia Institute of Technology 2011/04/12 1.
1 1 COMP5331: Knowledge Discovery and Data Mining Acknowledgement: Slides modified based on the slides provided by Lawrence Page, Sergey Brin, Rajeev Motwani.
By R. O. Nanthini and R. Jayakumar.  tools used on the web to find the required information  Akeredolu officially described the Web as “a wide- area.
Exploring Traversal Strategy for Web Forum Crawling Yida Wang, Jiang-Ming Yang, Wei Lai, Rui Cai Microsoft Research Asia, Beijing SIGIR
1 Random Walks on the Click Graph Nick Craswell and Martin Szummer Microsoft Research Cambridge SIGIR 2007.
1 e-Resources on Social Sciences: Scopus. 2 Why Scopus?  A comprehensive abstract and citation database of peer-reviewed literature and quality web sources.
What to Look for in Pay Per Click Management Published by:
Accurately Interpreting Clickthrough Data as Implicit Feedback
Weighted Graphs and traveling Salesperson problem
Subaru brand image award sponsorship opportunity
Redraw these graphs so that none of the line intersect except at the vertices B C D E F G H.
The Binomial Theorem.
Information Retrieval and Web Search
PJ SEO Specialists WordPress Web Development and SEO.
Compact Query Term Selection Using Topically Related Text
Search Pages and Results
Sergey Brin, lawrence Page, The anatomy of a large scale hypertextual web search Engine Rogier Brussee ICI
Web Information retrieval (Web IR)
Movie Recommendation System
A graphing calculator is required for some problems or parts of problems 2000.
Agenda What is SEO ? How Do Search Engines Work? Measuring SEO success ? On Page SEO – Basic Practices? Technical SEO - Source Code. Off Page SEO – Social.
Probabilistic Databases
Web Information retrieval (Web IR)
RANDOM NUMBERS SET # 1:
Presentation transcript:

Millions of Databases: Which are Trustworthy and Relevant? Trust and Profit Sensitive Ranking for On-line Ads and Web Databases Raju Balakrishnan, advised by Subbarao Kambhampati rajub@asu.edu rao@asu.edu Optimal Ad-Ranking Considering Mutual-Influences Comparison of Placement Strategies Ad Placement Strategies Sort by Bid x Relevance Sort by Bid Ads are Considered in Isolation, Ignoring Mutual influences. We Consider Ads as a Set, and ranking is based on User’s Browsing Model Three Manifestations of Mutual Influences on an ad are Similar ads placed above Reduces user’s residual relevance of the ad Relevance of other ads placed above User may click on ads above and may not view the ad Abandonment probability of other ads placed above User may abandon search and not view the ad Expected Profit Optimal Ranking 45.7% 35.9% THEOREM: Optimal Ad Placement Considering Similarities between the ads is NP-Hard SourceRank: Trust and Relevance based Ranking of Web Databases for the Deep Web Millions of Databases: Which are Trustworthy and Relevant? Evaluated in TEL-8 movies and books web databases ( 22 each). Agreement Graph Agreement Implies Trust & Relevance! Top-5 Precision-Movies SourceRank is calculated as the stationary visit probability of a weighted random walk on the database vertex in agreement graph. Coverage of a Source is the sum of the relevances of the tuples Trustworthiness is evaluated as the decrease in ranks of corrupted sources. Combine Coverage and SourceRank http://rakaposhi.eas.asu.edu/scuba