Lecture 16: Web search/Crawling/Link Analysis

Slides:



Advertisements
Similar presentations
CS276 Information Retrieval and Web Search
Advertisements

Near-Duplicates Detection
Latent Semantic Indexing (mapping onto a smaller space of latent concepts) Paolo Ferragina Dipartimento di Informatica Università di Pisa Reading 18.
TrustRank Algorithm Srđan Luković 2010/3482
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 16: Web search basics.

MMDS Secs Slides adapted from: J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, October.
@ Carnegie Mellon Databases User-Centric Web Crawling Sandeep Pandey & Christopher Olston Carnegie Mellon University.
CS345 Data Mining Web Spam Detection. Economic considerations  Search has become the default gateway to the web  Very high premium to appear on the.
15-853Page :Algorithms in the Real World Indexing and Searching III (well actually II) – Link Analysis – Near duplicate removal.
Near Duplicate Detection
Finding Similar Items.
Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 19: Web Search 1.
Finding Near Duplicates (Adapted from slides and material from Rajeev Motwani and Jeff Ullman)
CS246 Link-Based Ranking. Problems of TFIDF Vector  Works well on small controlled corpus, but not on the Web  Top result for “American Airlines” query:
1 SOCIAL BOOKMARKING 101. HIBA KHALID BILAL SAEED KHAN FARID ALIANI ASKARI HASAN SOCIAL BOOKMARKING.

Web search basics.
ITCS 6265 Information Retrieval and Web Mining Lecture 10: Web search basics.
Search Engine Optimization (SEO) Week 07 Dynamic Web TCNJ Jean Chu.
Λ14 Διαδικτυακά Κοινωνικά Δίκτυα και Μέσα
Detecting Near-Duplicates for Web Crawling Manku, Jain, Sarma
Adversarial Information Retrieval The Manipulation of Web Content.
COMP4210 Information Retrieval and Search Engines Lecture 7: Web search basics.
Promotion & Cataloguing AGCJ 407 Web Authoring in Agricultural Communications.
CS276 Lecture 16 Web characteristics II: Web size measurement Near-duplicate detection.
Web Search Module 6 INST 734 Doug Oard. Agenda The Web  Crawling Web search.
Introduction to Information Retrieval Introduction to Information Retrieval CS276 Information Retrieval and Web Search Chris Manning and Pandu Nayak Crawling.
Searching the Web From Chris Manning’s slides for Lecture 8: IIR Chapter 19.
Overview What is a Web search engine History Popular Web search engines How Web search engines work Problems.
Search Engine Marketing Gay, Charlesworth & Esen Chapter 6.
When Experts Agree: Using Non-Affiliated Experts To Rank Popular Topics Meital Aizen.
| 1 › Gertjan van Noord2014 Search Engines Lecture 7: relevance feedback & query reformulation.
윤언근 DataMining lab.  The Web has grown exponentially in size but this growth has not been isolated to good-quality pages.  spamming and.
Brief (non-technical) history Full-text index search engines Altavista, Excite, Infoseek, Inktomi, ca Taxonomies populated with web page Yahoo.
Search engines are the key to finding specific information on the vast expanse of the World Wide Web. Without sophisticated search engines, it would be.
Improving Cloaking Detection Using Search Query Popularity and Monetizability Kumar Chellapilla and David M Chickering Live Labs, Microsoft.
Search Engines: The players and the field The mechanics of a typical search. The search engine wars. Statistics from search engine logs. The architecture.
Search Engine Marketing SEM = Search Engine Marketing SEO = Search Engine Optimization optimizing (altering/changing) your page in order to get a higher.
Web Search Module 6 INST 734 Doug Oard. Agenda The Web Crawling  Web search.
Web Search Basics (Modified from Stanford CS276 Class Lecture 13: Web search basics)
Document duplication (exact or approximate) Paolo Ferragina Dipartimento di Informatica Università di Pisa Slides only!
DATA MINING LECTURE 6 Sketching, Min-Hashing, Locality Sensitive Hashing.
SEO Friendly Website Building a visually stunning website is not enough to ensure any success for your online presence.
Information Retrieval (9) Prof. Dragomir R. Radev
Search Engine and Optimization 1. Introduction to Web Search Engines 2.
Introduction to Information Retrieval (Manning, Raghavan, Schutze) Chapter 19 Web search basics.
Web Spam Taxonomy Zoltán Gyöngyi, Hector Garcia-Molina Stanford Digital Library Technologies Project, 2004 presented by Lorenzo Marcon 1/25.
WEB SEARCH BASICS By K.KARTHIKEYAN. Web search basics The Web Ad indexes Web spider Indexer Indexes Search User Sec
Search Engine Optimization
CS276 Information Retrieval and Web Search
Search Engine Optimization
CS276A Text Information Retrieval, Mining, and Exploitation
Modified by Dongwon Lee from slides by
CCT356: Online Advertising and Marketing
WEB SPAM.
15-499:Algorithms and Applications
Near Duplicate Detection
Crawler (AKA Spider) (AKA Robot) (AKA Bot)
Finding Near-Duplicate Web Pages: A Large-Scale Evaluation of Algorithms By Monika Henzinger Presented.
IS 360 Web Promotion.
PageRank Random Surfers on the Web Transition Matrix of the Web Dead Ends and Spider Traps Topic-Specific PageRank Jeffrey D. Ullman Stanford University.
BTEC NCF Dip in Comp - Unit 15 Website Development Lesson 04 – Search Engine Optimisation Mr C Johnston.
Information Retrieval Christopher Manning and Prabhakar Raghavan
Information Retrieval on the World Wide Web
Agenda What is SEO ? How Do Search Engines Work? Measuring SEO success ? On Page SEO – Basic Practices? Technical SEO - Source Code. Off Page SEO – Social.
Web Search Engines.
Junghoo “John” Cho UCLA
Minwise Hashing and Efficient Search
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Presentation transcript:

Lecture 16: Web search/Crawling/Link Analysis

Brief (non-technical) history Early keyword-based engines ca. 1995-1997 Altavista, Excite, Infoseek, Inktomi, Lycos Paid search ranking: Goto (morphed into Overture.com  Yahoo!) Your search ranking depended on how much you paid Auction for keywords: casino was expensive!

Brief (non-technical) history 1998+: Link-based ranking pioneered by Google Blew away all early engines save Inktomi Great user experience in search of a business model Meanwhile Goto/Overture’s annual revenues were nearing $1 billion Result: Google added paid search “ads” to the side, independent of search results Yahoo followed suit, acquiring Overture (for paid placement) and Inktomi (for search) 2005+: Google gains search share, dominating in Europe and very strong in North America 2009: Yahoo! and Microsoft propose combined paid search offering

Paid Search Ads Algorithmic results.

Web search basics User The Web Indexer Indexes Ad indexes Web spider Sec. 19.4.1 Web search basics User The Web Web spider Indexer Search Indexes Ad indexes

User Needs Need [Brod02, RL04] Sec. 19.4.1 User Needs Need [Brod02, RL04] Informational – want to learn about something (~40% / 65%) Navigational – want to go to that page (~25% / 15%) Transactional – want to do something (web-mediated) (~35% / 20%) Access a service Downloads Shop Gray areas Find a good hub Exploratory search “see what’s there” Low hemoglobin United Airlines Seattle weather Mars surface images Canon S410 Car rental Brasil

How far do people look for results? (Source: iprospect.com WhitePaper_2006_SearchEngineUserBehavior.pdf)

SPAM (SEARCH ENGINE OPTIMIZATION)

Simplest forms First generation engines relied heavily on tf/idf Sec. 19.2.2 Simplest forms First generation engines relied heavily on tf/idf The top-ranked pages for the query maui resort were the ones containing the most maui’s and resort’s SEOs responded with dense repetitions of chosen terms e.g., maui resort maui resort maui resort Often, the repetitions would be in the same color as the background of the web page Repeated terms got indexed by crawlers But not visible to humans on browsers Pure word density cannot be trusted as an IR signal

Cloaking Serve fake content to search engine spider Sec. 19.2.2 Cloaking Serve fake content to search engine spider DNS cloaking: Switch IP address. Impersonate Is this a Search Engine spider? N Y SPAM Real Doc Cloaking

More spam techniques Doorway pages Link spamming Robots Sec. 19.2.2 More spam techniques Doorway pages Pages optimized for a single keyword that re-direct to the real target page Link spamming Mutual admiration societies, hidden links, awards – more on these later Domain flooding: numerous domains that point or re-direct to a target page Robots Fake query stream – rank checking programs “Curve-fit” ranking programs of search engines Millions of submissions via Add-Url

The war against spam Quality signals - Prefer authoritative pages based on: Votes from authors (linkage signals) Votes from users (usage signals) Policing of URL submissions Anti robot test Limits on meta-keywords Robust link analysis Ignore statistically implausible linkage (or text) Use link analysis to detect spammers (guilt by association) Spam recognition by machine learning Training set based on known spam Family friendly filters Linguistic analysis, general classification techniques, etc. For images: flesh tone detectors, source text analysis, etc. Editorial intervention Blacklists Top queries audited Complaints addressed Suspect pattern detection

SIZE OF THE WEB

What is the size of the web ? Sec. 19.5 What is the size of the web ? Issues The web is really infinite Dynamic content, e.g., calendars Soft 404: www.yahoo.com/<anything> is a valid page Static web contains syntactic duplication, mostly due to mirroring (~30%) Some servers are seldom connected Who cares? Media, and consequently the user Engine design Engine crawl policy. Impact on recall.

What can we attempt to measure? Sec. 19.5 What can we attempt to measure? The relative sizes of search engines The notion of a page being indexed is still reasonably well defined. Already there are problems Document extension: e.g., engines index pages not yet crawled, by indexing anchortext. Document restriction: All engines restrict what is indexed (first n words, only relevant words, etc.)

Relative Size from Overlap Given two engines A and B Sec. 19.5 Relative Size from Overlap Given two engines A and B A Ç B Sample URLs randomly from A Check if contained in B and vice versa A Ç B = (1/2) * Size A A Ç B = (1/6) * Size B (1/2)*Size A = (1/6)*Size B \ Size A / Size B = (1/6)/(1/2) = 1/3 Each test involves: (i) Sampling (ii) Checking

Sec. 19.6 DUPLICATE DETECTION

Duplicate documents The web is full of duplicated content Sec. 19.6 Duplicate documents The web is full of duplicated content Strict duplicate detection = exact match Not as common But many, many cases of near duplicates E.g., last-modified date the only difference between two copies of a page

Duplicate/Near-Duplicate Detection Sec. 19.6 Duplication: Exact match can be detected with fingerprints Near-Duplication: Approximate match Overview Compute syntactic similarity with an edit-distance measure Use similarity threshold to detect near-duplicates E.g., Similarity > 80% => Documents are “near duplicates” Not transitive though sometimes used transitively

Computing Similarity Features: Sec. 19.6 Computing Similarity Features: Segments of a document (natural or artificial breakpoints) Shingles (Word N-Grams) a rose is a rose is a rose → a_rose_is_a rose_is_a_rose is_a_rose_is Similarity Measure between two docs (= sets of shingles) Jaccard coefficient: Size_of_Intersection / Size_of_Union

Shingles + Set Intersection Computing exact set intersection of shingles between all pairs of documents is expensive/intractable Approximate using a cleverly chosen subset of shingles from each (a sketch) Estimate (size_of_intersection / size_of_union) based on a short sketch Doc A Shingle set A Sketch A Doc B Shingle set B Sketch B Jaccard

Sec. 19.6 Sketch of a document Create a “sketch vector” (of size ~200) for each document Documents that share ≥ t (say 80%) corresponding vector elements are near duplicates For doc D, sketchD[ i ] is as follows: Let f map all shingles in the universe to 0..2m-1 (e.g., f = fingerprinting) Let pi be a random permutation on 0..2m-1 Pick MIN {pi(f(s))} over all shingles s in D

Computing Sketch[i] for Doc1 Sec. 19.6 Computing Sketch[i] for Doc1 Document 1 264 Start with 64-bit f(shingles) Permute on the number line with pi Pick the min value 264 264 264

Test if Doc1.Sketch[i] = Doc2.Sketch[i] Sec. 19.6 Test if Doc1.Sketch[i] = Doc2.Sketch[i] Document 2 Document 1 264 264 264 264 264 264 A B 264 264 Are these equal? Test for 200 random permutations: p1, p2,… p200

However… Document 2 Document 1 264 Sec. 19.6 However… Document 1 Document 2 264 A B A = B iff the shingle with the MIN value in the union of Doc1 and Doc2 is common to both (i.e., lies in the intersection) Claim: This happens with probability Size_of_intersection / Size_of_union Why?

Set Similarity of sets Ci , Cj Sec. 19.6 Set Similarity of sets Ci , Cj View sets as columns of a matrix A; one row for each element in the universe. aij = 1 indicates presence of item i in set j Example C1 C2 0 1 1 0 1 1 Jaccard(C1,C2) = 2/5 = 0.4 0 0 1 1 0 1

Key Observation For columns Ci, Cj, four types of rows Sec. 19.6 Key Observation For columns Ci, Cj, four types of rows Ci Cj A 1 1 B 1 0 C 0 1 D 0 0 Overload notation: A = # of rows of type A Claim

“Min” Hashing Randomly permute rows Sec. 19.6 “Min” Hashing Randomly permute rows Hash h(Ci) = index of first row with 1 in column Ci Surprising Property Why? Both are A/(A+B+C) Look down columns Ci, Cj until first non-Type-D row h(Ci) = h(Cj)  type A row

Min-Hash sketches Pick P random row permutations MinHash sketch Sec. 19.6 Min-Hash sketches Pick P random row permutations MinHash sketch SketchD = list of P indexes of first rows with 1 in column C Similarity of signatures Let sim[sketch(Ci),sketch(Cj)] = fraction of permutations where MinHash values agree Observe E[sim(sketch(Ci),sketch(Cj))] = Jaccard(Ci,Cj)

Example Signatures S1 S2 S3 Perm 1 = (12345) 1 2 1 Sec. 19.6 Example Signatures S1 S2 S3 Perm 1 = (12345) 1 2 1 Perm 2 = (54321) 4 5 4 Perm 3 = (34512) 3 5 4 C1 C2 C3 R1 1 0 1 R2 0 1 1 R3 1 0 0 R4 1 0 1 R5 0 1 0 Similarities 1-2 1-3 2-3 Col-Col 0.00 0.50 0.25 Sig-Sig 0.00 0.67 0.00

More resources IIR Chapter 19