Information Retrieval Instructor Marina Gavrilova Outline Information Retrieval IR vs DBMS Boolean Text Search Text Indexes Simple relational text index Example of inverted file Computing ID: 389742
Download Presentation The PPT/PDF document "Introduction to" is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.
Slide1
Introduction to Information Retrieval
Instructor : Marina GavrilovaSlide2
Outline
Information Retrieval
IR
vs
DBMS
Boolean Text Search
Text Indexes
Simple relational text index
Example of inverted file
Computing
Relevance
Vector Space Model
Text
Clustering
Probabilistic Models and Ranking Principles
Iterative query refinement - Racchio Model
Query Modification
Collaborative Filtering and Ringo Collaborative Filtering
ConclusionsSlide3
Goal Goal of this lecture is to introduce you to Informational retrieval and how it differentiates from DBMS.
Then
we will discuss how vector space model and Text Clustering help in computing relevance and similarity between documents. Slide4
Information Retrieval
A research field traditionally separate from Databases
Goes back to IBM, Rand and Lockheed in the 50’s
G. Salton at Cornell in the 60’s
Lots of research since then
Products traditionally separate
Originally, document management systems for libraries, government, law, etc.
Gained prominence in recent years due to web searchSlide5
IR vs. DBMS
Seem like very different beasts:
Both
support queries over large datasets, use indexing.
In practice, you currently have to choose between the two.
IR
DBMS
Imprecise Semantics
Precise Semantics
Keyword search
SQL
Unstructured data format
Structured data
Read-Mostly. Add docs occasionally
Expect reasonable number of updates
Page through
top
k
results
Generate full answerSlide6
IR’s “Bag of Words” Model
Typical IR data model:
Each document is just a bag (multiset) of words (“terms”)
Detail 1: “Stop Words”
Certain words are considered irrelevant and not placed in the bag
e.g., “the”
e.g., HTML tags like <H1>
Detail 2: “Stemming” and other content analysis
Using English-specific rules, convert words to their basic form
e.g., “surfing”, “surfed” --> “surf”Slide7
Boolean Text Search
Find all documents that match a Boolean containment expression:
“Windows”
AND (“Glass” OR “Door”)
AND NOT “Microsoft”
Note:
Query terms are also filtered via stemming and stop words.
When web search engines say “10,000 documents found”, that’s the Boolean search result size (subject to a common “max # returned’ cutoff).Slide8
Text “Indexes”When IR
says
“text index”…
Usually mean more than what DB people mean
Both
“tables” and indexes
Really a logical schema (i.e., tables)
With a physical schema (i.e., indexes)
Usually not stored in a
DBMSSlide9
A Simple Relational Text Index
Create and populate a table
InvertedFile(term
string,
docURL
string
)
Build a B+-tree or Hash index on
InvertedFile.term
<Key, list of URLs> as entries in index critical here for efficient storage!! Fancy list compression possibleNote
: URL instead of RID, the web is your “heap file”!
Can also
cache
pages and use RIDs
This is often called an “inverted file” or “inverted index”
Maps from
words -> docs
Can now do single-word text search queries!Slide10
An Inverted File
Search for
“databases
”Slide11
Computing Relevance, Similarity: The Vector Space ModelSlide12
Document VectorsDocuments are represented as “bags of words”Represented as vectors when used computationally
A vector is like an array of floating point
Has direction and magnitude
Each vector holds a place for
every
term in the collection
Therefore, most vectors are sparseSlide13
Document Vectors:One location for each word.
nova galaxy heat h’wood film role diet fur
10 5 3
5 10
10 8 7
9 10 5
10 10
9 10
5 7 9
6 10 2 8
7 5 1 3
A
B
C
D
E
F
G
H
I
“Nova” occurs 10 times in text A
“Galaxy” occurs 5 times in text A
“Heat” occurs 3 times in text A
(Blank means 0 occurrences.)Slide14
Document Vectors
nova galaxy heat h’wood film role diet fur
10 5 3
5 10
10 8 7
9 10 5
10 10
9 10
5 7 9
6 10 2 8 7 5 1 3
A
B
C
D
E
F
G
H
I
Document idsSlide15
We Can Plot the Vectors
Star
Diet
Doc about astronomy
Doc about movie stars
Doc about mammal behavior
Assumption:
Documents that are “close” in space are similar. Slide16
Vector Space Model
Documents are represented as
vectors
in term space
Terms are usually stems
Documents represented by binary vectors of terms
Queries represented the same as documents
A vector distance measure between the query and documents is used to rank retrieved documents
Query and Document similarity is based on length and direction of their vectors
Vector operations to capture boolean query conditionsTerms in a vector can be “weighted” in many waysSlide17
Vector Space Documentsand Queries
D
1
D
2
D
3
D
4
D
5
D
6
D
7
D
8
D
9
D
10
D
11
t
2
t
3
t
1
Boolean term combinations
Q is a query – also represented
as a vectorSlide18
Assigning Weights to Terms
Binary Weights
Raw term frequency
Want
to weight terms highly if they are
frequent in relevant documents … BUT
infrequent in the collection as a wholeSlide19
Binary Weights
Only the presence (1) or absence (0) of a term is included in the vectorSlide20
Raw Term WeightsThe frequency of occurrence for the term in each document is included in the vectorSlide21
TF x IDF Normalization
Normalize the term weights (so longer documents are not unfairly given more weight)
The longer the document, the more likely it is for a given term to appear in it, and the more often a given term is likely to appear in it. So, we want to reduce the importance attached to a term appearing in a document based on the length of the document.Slide22
Pair-wise Document Similarity
nova galaxy heat h’wood film role diet fur
1 3 1
5 2
2 1 5
4 1
A
B
C
D
How to compute document similarity?Slide23
Pair-wise Document Similarity
nova galaxy heat h’wood film role diet fur
1 3 1
5 2
2 1 5
4 1
A
B
C
DSlide24
Pair-wise Document Similarity(cosine normalization)Slide25
Text ClusteringFinds overall similarities among groups of documentsFinds overall similarities among groups of tokens
Picks out some themes, ignores othersSlide26
Text Clustering
Clustering is
“The
art
of finding groups in data.”
-- Kaufmann and Rousseeu
Term 1
Term 2Slide27
Problems with Vector SpaceThere is no real theoretical basis for the assumption of a term space
It is more for visualization than having any real basis
Most similarity measures work about the same
Terms are not really orthogonal dimensions
Terms are not independent of all other terms; remember our discussion of correlated terms in textSlide28
Probabilistic ModelsRigorous formal model attempts to predict the probability that a given document will be relevant to a given query
Ranks retrieved documents according to this probability of relevance
(Probability Ranking Principle)
Relies on accurate estimates of probabilitiesSlide29
Probability Ranking PrincipleIf a reference retrieval system’s response to each request is a ranking of the documents in the collections in the order of decreasing probability of usefulness to the user who submitted the request, where the probabilities are estimated as accurately as possible on the basis of whatever data has been made available to the system for this purpose, then the overall effectiveness of the system to its users will be the best that is obtainable on the basis of that data.
Stephen E. Robertson,
J. Documentation
1977Slide30
Query Modification
Problem:
How can we reformulate the query to help a user who is trying several searches to get at the same information?
Thesaurus expansion:
Suggest terms similar to query terms
Relevance feedback:
Suggest terms (and documents) similar to retrieved documents that have been judged to be relevantSlide31
Relevance Feedback
Main Idea:
Modify existing query based on relevance judgements
Extract terms from relevant documents and add them to the query
AND/OR re-weight the terms already in the query
There are many variations:
Usually
positive weights
for terms from
relevant docsSometimes negative weights for terms from non-relevant docs
Users, or the system, guide this process by selecting terms from an automatically-generated list.
Slide32
Rocchio MethodRocchio automatically
Re-weights terms
Adds in new terms (from relevant docs)
have to be careful when using negative terms
Rocchio is
not
a machine learning algorithmSlide33
Rocchio MethodSlide34
Alternative Notions of Relevance Feedback
Find people whose taste is “similar” to yours.
Will you like what they like?
Follow a user’s actions in the background.
Can this be used to predict what the user will want to see next?
Track what lots of people are doing.
Does this implicitly indicate what they think is good and not good?Slide35
Collaborative Filtering (Social Filtering)
If Pam liked the paper, I’ll like the paper
If you liked Star Wars, you’ll like Independence Day
Rating based on ratings of similar people
Ignores text, so also works on sound, pictures etc.
But: Initial users can bias ratings of future usersSlide36
Users rate items from
like
to
dislike
7 = like; 4 = ambivalent; 1 = dislike
A normal distribution; the extremes are what matter
Nearest Neighbors Strategy:
Find similar users and predicted (weighted) average of user ratings
Pearson Algorithm:
Weight by degree of correlation between user U and user J1 means similar, 0 means no correlation, -1 dissimilarWorks better to compare against the ambivalent rating (4), rather than the individual’s average score
Ringo Collaborative FilteringSlide37
Computing Relevance
Relevance calculation involves how often search terms appear in doc, and how often they appear in collection:
More search terms found in doc
doc is more relevant
Greater importance attached to finding
rare
terms
Doing this efficiently in current SQL engines is not easy:
“Relevance of a doc wrt a search term” is a function that is called once per doc the term appears in (docs found via inv. index):
For efficient fn computation, for each term, we can store the # times it appears in each doc, as well as the # docs it appears in.Must also sort retrieved docs by their relevance value.Also, think about Boolean operators (if the search has multiple terms) and how they affect the relevance computation!
An object-relational or object-oriented DBMS with good support for function calls is better, but you still have long execution path-lengths compared to optimized search engines.Slide38
Updates and Text Search
Text search engines are designed to be query-mostly:
Deletes and modifications are rare
Can postpone updates (nobody notices, no transactions!)
Updates done in batch (rebuild the index)
Can’t afford to go off-line for an update?
Create a 2nd index on a separate machine
Replace the 1st index with the 2nd!
So no concurrency control problems
Can compress to search-friendly, update-unfriendly formatMain reason why text search engines and DBMSs are usually separate products.Also, text-search engines tune that one SQL query to death!Slide39
{
DBMS vs. Search Engine Architecture
The Access Method
Buffer Management
Disk Space Management
OS
“The Query”
Search String Modifier
Simple
DBMS
}
Ranking Algorithm
Query Optimization
and Execution
Relational Operators
Files and Access Methods
Buffer Management
Disk Space Management
Concurrency
and
Recovery
Needed
DBMS
Search EngineSlide40
IR vs. DBMS Revisited
Semantic Guarantees
DBMS guarantees transactional semantics
If inserting Xact commits, a later query
will see
the update
Handles multiple concurrent updates correctly
IR systems do not do this; nobody notices!
Postpone insertions until convenient
No model of correct concurrencyData Modeling & Query ComplexityDBMS supports any schema & queries
Requires you to define schema
Complex query language hard to learn
IR supports only one schema & query
No schema design required (unstructured text)
Trivial to learn query languageSlide41
Lots More in IR …
How to “rank” the output? I.e., how to compute relevance of each result item w.r.t. the query?
Doing this well / efficiently is hard!
Other ways to help users paw through the output?
Document “clustering”, document visualization
How to take advantage of hyperlinks?
Really cute tricks here!
How to use compression for better I/O performance?
E.g., making RID lists smaller
Try to make things fit in RAM!How to deal with synonyms, misspelling, abbreviations?
How to write a good web crawler?Slide42
Summary
First we studied difference between Information Retrieval and DBMS . Then we disused on two type of searches (Boolean and Text based search) used in DBMS .
In addition, we learned how we can compute relevance between documents based on words using Vector Space Model and how text clustering can be used to find similarity between documents and in the end we discussed
Racchio Model for iterative query refinement. Slide43
SummaryIR relies on computing distance between documentsTerms can be weighted and distances normalized
IR can utilize clustering, adaptive query updates and elements of learning to perform document retrieval / response to query better
Idea is to use not only similarity, but dissimilarity measures to
compare documents.