KDD 2011 Research Poster Content - Driven Trust Propagation Framwork V. G. Vinod Vydiswaran, ChengXiang Zhai, and Dan Roth University of Illinois at Urbana-Champaign.

Slides:



Advertisements
Similar presentations
1 Opinion Summarization Using Entity Features and Probabilistic Sentence Coherence Optimization (UIUC at TAC 2008 Opinion Summarization Pilot) Nov 19,
Advertisements

Individual Campaign Contributions and Candidate Ideology ARTICLE BY MICHAEL J. ENSLEY PRESENTATION BY WILLIAM JERGINS.
Trust and Profit Sensitive Ranking for Web Databases and On-line Advertisements Raju Balakrishnan (Arizona State University)
Opinion Spam and Analysis Nitin Jindal and Bing Liu Department of Computer Science University of Illinois at Chicago.
Jean-Eudes Ranvier 17/05/2015Planet Data - Madrid Trustworthiness assessment (on web pages) Task 3.3.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
Compare&Contrast: Using the Web to Discover Comparable Cases for News Stories Presenter: Aravind Krishna Kalavagattu.
Analyzing Sentiment in a Large Set of Web Data while Accounting for Negation AWIC 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam.
USING STUDENT OUTCOMES WHEN INTEGRATING INFORMATION LITERACY SKILLS INTO COURSES Information Literacy Department Asa H. Gordon Library Savannah State University.
Random Thoughts 2012 (COMP 066) Jan-Michael Frahm Jared Heinly source: fivethirtyeight.com.
Signatures As Threats to Privacy Brian Neil Levine Assistant Professor Dept. of Computer Science UMass Amherst.
Social Networking and On-Line Communities: Classification and Research Trends Maria Ioannidou, Eugenia Raptotasiou, Ioannis Anagnostopoulos.
Advocacy.
Cluster based fact finders Manish Gupta, Yizhou Sun, Jiawei Han Feb 10, 2011.
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Can you believe what you read online?: Modeling and Predicting Trustworthiness of Online Textual Information V.G.Vinod Vydiswaran Department of Computer.
2008 © ChengXiang Zhai Dragon Star Lecture at Beijing University, June 21-30, Prepare Yourself for IR Research ChengXiang Zhai Department of Computer.
Exploiting Wikipedia as External Knowledge for Document Clustering Sakyasingha Dasgupta, Pradeep Ghosh Data Mining and Exploration-Presentation School.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
CS598CXZ (CS510) Advanced Topics in Information Retrieval (Fall 2014) Instructor: ChengXiang (“Cheng”) Zhai 1 Teaching Assistants: Xueqing Liu, Yinan Zhang.
BiasTrust: Teaching Biased Users About Controversial Topics V.G.Vinod Vydiswaran, ChengXiang Zhai, Dan Roth University of Illinois at Urbana-Champaign.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
SIRS DISCOVERER BY PROQUEST. Overview Sources and articles are selected for their educational content, reliability, relevance, interest, age- appropriateness,
PAUL ALEXANDRU CHIRITA STEFANIA COSTACHE SIEGFRIED HANDSCHUH WOLFGANG NEJDL 1* L3S RESEARCH CENTER 2* NATIONAL UNIVERSITY OF IRELAND PROCEEDINGS OF THE.
1 Mining User Behavior Mining User Behavior Eugene Agichtein Mathematics & Computer Science Emory University.
2008 © ChengXiang Zhai Dragon Star Lecture at Beijing University, June 21-30, Frame an IR Research Problem and Form Hypotheses ChengXiang Zhai Department.
Automatic Detection of Tags for Political Blogs Khairun-nisa Hassanali Vasileios Hatzivassiloglou The University.
ANSWERING APUSH ESSAY QUESTIONS (FRQ). Essay Prompt All college-level essay test answers are written in response to an essay “prompt.” All college-level.
THE DEAD SEA & DENSITY CAN NEWS ARTICLES BE TRUSTED TO PROVIDE ACCURATE INFORMATION?
A General Optimization Framework for Smoothing Language Models on Graph Structures Qiaozhu Mei, Duo Zhang, ChengXiang Zhai University of Illinois at Urbana-Champaign.
Learning from Multi-topic Web Documents for Contextual Advertisement KDD 2008.
Context-Sensitive Information Retrieval Using Implicit Feedback Xuehua Shen : department of Computer Science University of Illinois at Urbana-Champaign.
Toward A Session-Based Search Engine Smitha Sriram, Xuehua Shen, ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Relevance Feedback: New Trends Derive global optimization methods: More computationally robust Consider the correlation between different attributes Incorporate.
University of Malta CSA3080: Lecture 6 © Chris Staff 1 of 20 CSA3080: Adaptive Hypertext Systems I Dr. Christopher Staff Department.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
1 Web-Page Summarization Using Clickthrough Data* JianTao Sun, Yuchang Lu Dept. of Computer Science TsingHua University Beijing , China Dou Shen,
Erasmus University Rotterdam Introduction Content-based news recommendation is traditionally performed using the cosine similarity and TF-IDF weighting.
BiasTrust: Trusting credible information in presence of human bias V.G.Vinod Vydiswaran ChengXiang Zhai, Dan Roth Department of Computer Science University.
MOTIVATION AND CHALLENGE Big data Volume Velocity Variety Veracity Contributor Content Context Value 5 Vs of Big Data 3 Cs of Veracity.
Positional Relevance Model for Pseudo–Relevance Feedback Yuanhua Lv & ChengXiang Zhai Department of Computer Science, UIUC Presented by Bo Man 2014/11/18.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Blog Summarization We have built a blog summarization system to assist people in getting opinions from the blogs. After identifying topic-relevant sentences,
1 A Formal Study of Information Retrieval Heuristics Hui Fang, Tao Tao and ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
KDD 2011 Doctoral Session Modeling Trustworthiness of Online Content V. G. Vinod Vydiswaran Advisors: Prof.ChengXiang Zhai, Prof.Dan Roth University of.
Exploring in the Weblog Space by Detecting Informative and Affective Articles Xiaochuan Ni, Gui-Rong Xue, Xiao Ling, Yong Yu Shanghai Jiao-Tong University.
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
UWMS Data Mining Workshop Content Analysis: Automated Summarizing Prof. Marti Hearst SIMS 202, Lecture 16.
Multi-level Bootstrapping for Extracting Parallel Sentence from a Quasi-Comparable Corpus Pascale Fung and Percy Cheung Human Language Technology Center,
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Improving the performance of personal name disambiguation.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
Automatic Labeling of Multinomial Topic Models
Michael Bendersky, W. Bruce Croft Dept. of Computer Science Univ. of Massachusetts Amherst Amherst, MA SIGIR
Myth Busting Distinguish between factual statements and inferences. Evaluate scientifically related claims against available evidence. Reject pseudoscience.
The Development of a search engine & Comparison according to algorithms Sung-soo Kim The final report.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Automatic Labeling of Multinomial Topic Models Qiaozhu Mei, Xuehua Shen, and ChengXiang Zhai DAIS The Database and Information Systems Laboratory.
Text Information Management ChengXiang Zhai, Tao Tao, Xuehua Shen, Hui Fang, Azadeh Shakery, Jing Jiang.
ENHANCING CLUSTER LABELING USING WIKIPEDIA David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab SIGIR’09.
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
哈工大信息检索研究室 HITIR ’ s Update Summary at TAC2008 Extractive Content Selection Using Evolutionary Manifold-ranking and Spectral Clustering Reporter: Ph.d.
INFORMATION RETRIEVAL MEASUREMENT OF RELEVANCE EFFECTIVENESS 1Adrienn Skrop.
Unsupervised Sparse Vector Densification for Short Text Similarity
On Dataless Hierarchical Text Classification
John Lafferty, Chengxiang Zhai School of Computer Science
Unit 7: Gathering Facts: Multiple Accounts
Unit 7: Gathering Facts: Multiple Accounts
Modeling Trust and Influence in the Blogosphere using Link Polarity
Presentation transcript:

KDD 2011 Research Poster Content - Driven Trust Propagation Framwork V. G. Vinod Vydiswaran, ChengXiang Zhai, and Dan Roth University of Illinois at Urbana-Champaign Incorporating text in trust models Model parameters Can you trust news stories?  Even reputed sources make mistakes.  Not all claims made by a source is equally trustworthy.  Some claims are purposefully misleading.  How to verify free-text claims? Acknowledgments This research was supported by the Multimodal Information Access and Synthesis (MIAS) Center at the University of Illinois at Urbana-Champaign, part of CCICADA, a DHS Science and Technology Center of Excellence, and grants from the Army Research Laboratory under agreement W911NF Contact details Claim 1 Claim n Claim EvidenceClaims Sources Web sources Evidence passages Claim sentences  Incorporates semantics in trust computation using evidence.  Claims need not be structured tuples – they can be free-text sentences.  Framework does not assume that accurate Information Extraction is available.  A source can have different trust profile for different claims – not all claims from a source get equal weight. Advantages over traditional models Traditional two-layer fact-finder models Claim 1 Claim n Claim 2 … [Yin, et al., 2007; Pasternack & Roth, 2010]  Computed scores : Claim veracity : Evidence trust : Source trust  Influence factors : Evidence similarity : Relevance : Source - Evidence influence Iterative formulation #TopicRetrievalTwo-stage models Our model 1Healthcare Obama administration Bush administration Democratic policy Republican policy Immigration Gay rights Corruption Election reform WikiLeaks Average % Relative  There is a need to determine the truth value of a claim.  This value depends on its source as well as on evidence. Evidence documents influence each other and have different relevance to claims.  We developed a trust propagation framework that associates relevant evidence to claims and sources.  Global analysis of this data, taking into account relations between the stories, their relevance and their sources allows us to make progress in determining trustworthiness values over sources and claims.  Experiments with news trustworthiness show promising results on incorporating evidence in trustworthiness computation and improving “credibility” of retrieved results. Conclusions Data characteristics Experimental results D. Using trust model to boost evidence retrieval C. Does it depend on news genres? A. Computing trust scores and trusted sources for specific claim topics B. Finding trustworthy news sources and news reporters  Model brings credible documents to the top of the result list  Improvement in NDCG scores statistically significant.  Model helps bring out the disparity in credibility of reporting on specific topics  Model scores show influence of both popularity and average rating of articles.  Specific news sources appear to be trusted more for specific news genres.  23,164 news articles from 23 genres collected from Politics category of NewsTrust.org  All news articles were rated by human volunteers based on journalistic principles  Scored in the range [1,5], mean 3.70  Investigative reports most trusted (4.10), Advertisements least (2.43) Veracity of news reporting Trustworthiness of news stories Credibility of news sources