2007.04.03 - SLIDE 1IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00.

Slides:



Advertisements
Similar presentations
The Mathematics of Information Retrieval 11/21/2005 Presented by Jeremy Chapman, Grant Gelven and Ben Lakin.
Advertisements

Introduction to Information Retrieval Outline ❶ Latent semantic indexing ❷ Dimensionality reduction ❸ LSI in information retrieval 1.
Text Databases Text Types
Dimensionality Reduction PCA -- SVD
INF 141 IR METRICS LATENT SEMANTIC ANALYSIS AND INDEXING Crista Lopes.
Comparison of information retrieval techniques: Latent semantic indexing (LSI) and Concept indexing (CI) Jasminka Dobša Faculty of organization and informatics,
What is missing? Reasons that ideal effectiveness hard to achieve: 1. Users’ inability to describe queries precisely. 2. Document representation loses.
Latent Semantic Analysis
Hinrich Schütze and Christina Lioma
An Introduction to Latent Semantic Analysis
ISP 433/633 Week 10 Vocabulary Problem & Latent Semantic Indexing Partly based on G.Furnas SI503 slides.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
1 CS 430 / INFO 430 Information Retrieval Lecture 11 Latent Semantic Indexing Extending the Boolean Model.
1 Latent Semantic Indexing Jieping Ye Department of Computer Science & Engineering Arizona State University
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
CSM06 Information Retrieval Lecture 3: Text IR part 2 Dr Andrew Salway
SLIDE 1IS 240 – Spring 2009 Prof. Ray Larson University of California, Berkeley School of Information Principles of Information Retrieval.
SLIDE 1IS 240 – Spring 2011 Prof. Ray Larson University of California, Berkeley School of Information Principles of Information Retrieval.
Indexing by Latent Semantic Analysis Written by Deerwester, Dumais, Furnas, Landauer, and Harshman (1990) Reviewed by Cinthia Levy.
Information Retrieval in Text Part III Reference: Michael W. Berry and Murray Browne. Understanding Search Engines: Mathematical Modeling and Text Retrieval.
SLIDE 1IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00.
SLIDE 1IS 240 – Spring 2010 Prof. Ray Larson University of California, Berkeley School of Information Principles of Information Retrieval.
Singular Value Decomposition in Text Mining Ram Akella University of California Berkeley Silicon Valley Center/SC Lecture 4b February 9, 2011.
Indexing by Latent Semantic Analysis Scot Deerwester, Susan Dumais,George Furnas,Thomas Landauer, and Richard Harshman Presented by: Ashraf Khalil.
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
1/ 30. Problems for classical IR models Introduction & Background(LSI,SVD,..etc) Example Standard query method Analysis standard query method Seeking.
IR Models: Latent Semantic Analysis. IR Model Taxonomy Non-Overlapping Lists Proximal Nodes Structured Models U s e r T a s k Set Theoretic Fuzzy Extended.
Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 18: Latent Semantic Indexing 1.
Information Retrieval in Text Part II Reference: Michael W. Berry and Murray Browne. Understanding Search Engines: Mathematical Modeling and Text Retrieval.
SLIDE 1IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00.
E.G.M. PetrakisDimensionality Reduction1  Given N vectors in n dims, find the k most important axes to project them  k is user defined (k < n)  Applications:
Other IR Models Non-Overlapping Lists Proximal Nodes Structured Models Retrieval: Adhoc Filtering Browsing U s e r T a s k Classic Models boolean vector.
1 LSI (lecture 19) Using latent semantic analysis to improve access to textual information (Dumais et al, CHI-88) What’s the best source of info about.
1 CS 430 / INFO 430 Information Retrieval Lecture 9 Latent Semantic Indexing.
Software Engineering Laboratory, Department of Computer Science, Graduate School of Information Science and Technology, Osaka University Automatic Categorization.
Homework Define a loss function that compares two matrices (say mean square error) b = svd(bellcore) b2 = b$u[,1:2] %*% diag(b$d[1:2]) %*% t(b$v[,1:2])
1 Vector Space Model Rong Jin. 2 Basic Issues in A Retrieval Model How to represent text objects What similarity function should be used? How to refine.
CS246 Topic-Based Models. Motivation  Q: For query “car”, will a document with the word “automobile” be returned as a result under the TF-IDF vector.
Latent Semantic Analysis Hongning Wang VS model in practice Document and query are represented by term vectors – Terms are not necessarily orthogonal.
Latent Semantic Indexing Debapriyo Majumdar Information Retrieval – Spring 2015 Indian Statistical Institute Kolkata.
Indices Tomasz Bartoszewski. Inverted Index Search Construction Compression.
An Introduction to Latent Semantic Analysis. 2 Matrix Decompositions Definition: The factorization of a matrix M into two or more matrices M 1, M 2,…,
Latent Semantic Analysis Hongning Wang Recap: vector space model Represent both doc and query by concept vectors – Each concept defines one dimension.
CpSc 881: Information Retrieval. 2 Recall: Term-document matrix This matrix is the basis for computing the similarity between documents and queries. Today:
Latent Semantic Indexing: A probabilistic Analysis Christos Papadimitriou Prabhakar Raghavan, Hisao Tamaki, Santosh Vempala.
INFORMATION SEARCH Presenter: Pham Kim Son Saint Petersburg State University.
June 5, 2006University of Trento1 Latent Semantic Indexing for the Routing Problem Doctorate course “Web Information Retrieval” PhD Student Irina Veredina.
LATENT SEMANTIC INDEXING Hande Zırtıloğlu Levent Altunyurt.
SINGULAR VALUE DECOMPOSITION (SVD)
Alternative IR models DR.Yeni Herdiyeni, M.Kom STMIK ERESHA.
Gene Clustering by Latent Semantic Indexing of MEDLINE Abstracts Ramin Homayouni, Kevin Heinrich, Lai Wei, and Michael W. Berry University of Tennessee.
Latent Semantic Indexing and Probabilistic (Bayesian) Information Retrieval.
1 CS 430: Information Discovery Lecture 12 Latent Semantic Indexing.
LATENT SEMANTIC INDEXING BY SINGULAR VALUE DECOMPOSITION
1 Latent Concepts and the Number Orthogonal Factors in Latent Semantic Analysis Georges Dupret
Information Retrieval and Web Search Probabilistic IR and Alternative IR Models Rada Mihalcea (Some of the slides in this slide set come from a lecture.
1 CS 430: Information Discovery Lecture 11 Latent Semantic Indexing.
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
ITCS 6265 Information Retrieval & Web Mining Lecture 16 Latent semantic indexing Thanks to Thomas Hofmann for some slides.
Search Engine and Optimization 1. Agenda Indexing Algorithms Latent Semantic Indexing 2.
Plan for Today’s Lecture(s)
Best pTree organization? level-1 gives te, tf (term level)
Lecture 16: Filtering & TDT
LSI, SVD and Data Management
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
CS 430: Information Discovery
Packing to fewer dimensions
Information Retrieval and Web Search
Restructuring Sparse High Dimensional Data for Effective Retrieval
Latent Semantic Analysis
Presentation transcript:

SLIDE 1IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00 pm Spring Principles of Information Retrieval Lecture 18: Digital Libraries

SLIDE 2IS 240 – Spring 2007 Overview Review –Latent Semantic Indexing (LSI) Digital Libraries and IR

SLIDE 3IS 240 – Spring 2007 LSI Rationale The words that searchers use to describe the their information needs are often not the same words used by authors to describe the same information. I.e., index terms and user search terms often do NOT match –Synonymy –Polysemy Following examples from Deerwester, et al. Indexing by Latent Semantic Analysis. JASIS 41(6), pp , 1990

SLIDE 4IS 240 – Spring 2007 LSI Rationale Access Document Retrieval Information Theory Database Indexing Computer REL M D1 x x x x x R D2 x* x x* M D3 x x* x * R M Query: IDF in computer-based information lookup Only matching words are “information” and “computer” D1 is relevant, but has no words in the query…

SLIDE 5IS 240 – Spring 2007 LSI Rationale Problems of synonyms –If not specified by the user, will miss synonymous terms –Is automatic expansion from a thesaurus useful? –Are the semantics of the terms taken into account? Is there an underlying semantic model of terms and their usage in the database?

SLIDE 6IS 240 – Spring 2007 How LSI Works Start with a matrix of terms by documents Analyze the matrix using SVD to derive a particular “latent semantic structure model” Two-Mode factor analysis, unlike conventional factor analysis, permits an arbitrary rectangular matrix with different entities on the rows and columns –Such as Terms and Documents

SLIDE 7IS 240 – Spring 2007 How LSI Works The rectangular matrix is decomposed into three other matices of a special form by SVD –The resulting matrices contain “singular vectors” and “singular values” –The matrices show a breakdown of the original relationships into linearly independent components or factors –Many of these components are very small and can be ignored – leading to an approximate model that contains many fewer dimensions

SLIDE 8IS 240 – Spring 2007 How LSI Works Titles C1: Human machine interface for LAB ABC computer applications C2: A survey of user opinion of computer system response time C3: The EPS user interface management system C4: System and human system engineering testing of EPS C5: Relation of user-percieved response time to error measurement M1: The generation of random, binary, unordered trees M2: the intersection graph of paths in trees M3: Graph minors IV: Widths of trees and well-quasi-ordering M4: Graph minors: A survey Italicized words occur and multiple docs and are indexed

SLIDE 9IS 240 – Spring 2007 How LSI Works Terms Documents c1 c2 c3 c4 c5 m1 m2 m3 m4 Human Interface Computer User System Response Time EPS Survey Trees Graph Minors

SLIDE 10IS 240 – Spring 2007 How LSI Works Dimension 2 Dimension 1 11graph M2(10,11,12) 10 Tree 12 minor 9 survey M1(10) 7 time 3 computer 4 user 6 response 5 system 2 interface 1 human M4(9,11,12) M2(10,11) C2(3,4,5,6,7,9) C5(4,6,7) C1(1,2,3) C3(2,4,5,8) C4(1,5,8) Q(1,3) Blue dots are terms Documents are red squares Blue square is a query Dotted cone is cosine.9 from Query “Human Computer Interaction” -- even docs with no terms in common (c3 and c5) lie within cone. SVD to 2 dimensions

SLIDE 11IS 240 – Spring 2007 How LSI Works X = TSD’ T has orthogonal, unit-length columns (T’T = 1) D has orthogonal, unit-length columns (D’ D = 1) S 0 is the diagonal matrix of singular values t is the number of rows in X d is the number of columns in X m is the rank of X (<= min(t,d) k is the chosen number of dimensions in the reduced model (k <= m) XT = S D’ t x d t x k m x k k x d docs terms ^ ^

SLIDE 12IS 240 – Spring 2007 Comparisons in LSI Comparing two terms Comparing two documents Comparing a term and a document

SLIDE 13IS 240 – Spring 2007 Comparisons in LSI In the original matrix these amount to: –Comparing two rows –Comparing two columns –Examining a single cell in the table

SLIDE 14IS 240 – Spring 2007 Comparing Two Terms Dot product between the row vectors of X(hat) reflects the extent to which two terms have a similar pattern of occurrence across the set of documents

SLIDE 15IS 240 – Spring 2007 Comparing Two Documents The dot product between two column vectors of the matrix X(hat) which tells the extent to which two documents have a similar profile of terms

SLIDE 16IS 240 – Spring 2007 Comparing a term and a document Treat the query as a pseudo-document and calculate the cosine between the pseudo-document and the other documents

SLIDE 17IS 240 – Spring 2007 Use of LSI LSI has been tested and found to be “modestly effective” with traditional test collections. Permits compact storage/representation (vectors are typically elements instead of thousands) However, creating the compressed vectors from raw data is very compute-intensive and “search” is also

SLIDE 18IS 240 – Spring 2007 Today Digital Libraries and IR Image Retrieval in DL From paper presented at the 1999 ASIS Annual Meeting