2. Outlines
Concept
Practice
Challenge1. Conceptual Model
2. Retrieval Unit
3. Document
Representation
4. Information Needs
5. Indexing
6. Retrieval Functions
7. Evaluation
1. Search Engine
1. Cross Lingual IR
2. Big Data
3. Personalization
4. Domain Specific IR
5. ā¦ā¦
2
4. A Conceptual Model for IR
Documents
Document
Representation
Information Needs
Query
Retrieved
Documents
Indexing Formulation
Retrieval Function
Relevance Feedback
4
5. Definitions of IR
āInformation retrieval is a field concerned with the
structure, analysis, organization, storage, searching, and
retrieval of information.ā (Salton, 1968)
Information retrieval (IR) is finding material (usually
documents) of an unstructured nature (usually text) that
satisfies an information need from within large collections
(usually stored on computers). (Manning, 2008)
5
6. Document/Retrieval Unit
ā¦ Web pages, email, books, news stories, scholarly papers, text
messages, Wordā¢, Powerpointā¢, PDF, forum postings, patents,
etc.
ā¦ Retrieval unit can be
ā¦ Part of document, e.g. a paragraph, a slide, a page etc.
ā¦ In the form different structure, html, xml, text etc.
ā¦ In different sizes/length.
6
7. Document Representation
Full Text Representation
ā¦ Keep everything. Complete.
ā¦ Require huge resources. Too much may not be good.
Reduced (partial) Content Representation
ā¦ Remove not important contents e.g. stopwords.
ā¦ Standardization to reduce overlapped contents e.g. stemming.
ā¦ Retain only important contents, e.g. noun phrases, header etc.
7
8. Document Representation
Think of representation as some ways of storing the document.
Bag of Words Model
Store the words as the bag (multiset) of its words, disregarding grammar
and even word order.
Document 1: "The cat sat on the hat"
Document 2: "The dog ate the cat and the hat"
From these two documents, a word list is constructed:
{ the, cat, sat, on, hat, dog, ate, and }
The list has 8 distinct words.
Document 1: { 2, 1, 1, 1, 1, 0, 0, 0 }
Document 2 : { 3, 1, 0, 0, 1, 1, 1, 1}
8
9. Information Needs
Those things that you want Google to give you answer are
information needs.
Example of my search history
ā¦ Query: weka text classification
ā¦ Information Need: I want to find the tutorial about using weka for text
classification.
ā¦ Query: Dell i7 laptop
ā¦ Information Need: I want to find the information any dell laptop that runs on intel
i7 processor. Actually, I want to buy, so an online store would be relevant.
9
10. Information Needs
Normally, you are required to formulate your information needs into
some keywords, known as query.
Simple Query
ā¦ Few keywords or more.
Boolean Query
ā¦ āneural network AND speech recognitionā
Special Query
ā¦ 400 myr in usd
10
11. Retrieved Documents
From the original collection, a subset of documents are obtained.
What is the factor that determines what document to return?
Simple Term Matching Approach
1. Compare the terms in a document and query.
2. Compute āsimilarityā between each document in the collection and
the query based on the terms they have in common.
3. Sorting the document in order of decreasing similarity with the query.
4. The outputs are a ranked list and displayed to the user - the top ones
are more relevant as judged by the system.
11
12. Indexing
Convert documents into
representation or data structure to
improve the efficiency of retrieval.
To generate a set of useful terms
called indexes.
Why?
ā¦ Many variety of words used in texts,
but not all are important.
ā¦ Among the important words, some
are more contextually relevant.
Some basic processes involved
ā¦ Tokenization
ā¦ Stop Words Removal
ā¦ Stemming
ā¦ Phrases
ā¦ Inverted File
12
13. Indexing (Tokenization)
Convert a sequence of characters
into a sequence of tokens with
some basic meaning.
āThe cat chases the mouse.ā
āBigcorp's 2007 bi-annual report
showed profits rose 10%.ā
the
cat
chases
the
mouse
bigcorp
2007
bi
annual
report
showed
profits
rose
10%
13
14. Indexing (Tokenization)
Token can be single or multiple terms.
āSamsung Galaxy S7 Edge, redefines what a phone can do.ā
samsung galaxy s7 edge
redefines
what
a
phone
can
do
samsung
galaxy
s7
edge
redefines
what
a ā¦.
or
14
15. Indexing (Tokenization)
Common Issues
1. Capitalized words can have different meaning from lower case words
ā¦ Bush fires the officer. Query: Bush fire
ā¦ The bush fire lasted for 3 days. Query: bush fire
2. Apostrophes can be a part of a word, a part of a possessive, or just a
mistake
ā¦ rosie o'donnell, can't, don't, 80's, 1890's, men's straw hats, master's degree,
england's ten largest cities, shriner's
15
16. Indexing (Tokenization)
3. Numbers can be important, including decimals
ā¦ nokia 3250, top 10 courses, united 93, quicktime 6.5 pro, 92.3 the beat,
288358
4. Periods can occur in numbers, abbreviations, URLs, ends of sentences, and
other situations
ā¦ I.B.M., Ph.D., cs.umass.edu, F.E.A.R.
Note: tokenizing steps for queries must be identical to steps for documents
16
17. Indexing (Stopping)
Top 50 Words fromAP89 News
Collection
Recall,
Indexes should be useful term links
to a document.
Are the terms on the right figure
useful?
17
18. Indexing (Stopping)
Stopword list can be created from high-frequency words or based on
a standard list
Lists are customized for applications, domains, and even parts of
documents
ā¦ e.g., āclickā is a good stopword for anchor text
Best policy is to index all words in documents, make decisions about
which words to use at query time
18
19. Indexing (Stemming)
Many morphological variations of words
ā¦ inflectional (plurals, tenses)
ā¦ derivational (making verbs nouns etc.)
In most cases, these have the same or very similar meanings
Stemmers attempt to reduce morphological variations of words to a
common stem
ā¦ usually involves removing suffixes
Can be done at indexing time or as part of query processing (like
stopwords)
19
20. Indexing (Stemming)
Porter Stemmer
ā¦ Algorithmic stemmer used in
IR experiments since the 70s
ā¦ Consists of a series of rules
designed to the longest
possible suffix at each step
ā¦ Produces stems not words
ā¦ Example Step 1 (right figure)
20
22. Indexing (Phrases)
Recall, token, meaningful tokens are better indexes, e.g.
phrases.
Text processing issue ā how are phrases recognized?
Three possible approaches:
ā¦ Identify syntactic phrases using a part-of-speech (POS) tagger
ā¦ Use word n-grams
ā¦ Store word positions in indexes and use proximity operators in
queries
22
23. Indexing (Phrases)
POS taggers use statistical models of text to predict
syntactic tags of words
ā¦ Example tags:
ā¦ NN (singular noun), NNS (plural noun), VB (verb), VBD (verb, past tense), VBN (verb, past
participle), IN (preposition), JJ (adjective), CC (conjunction, e.g., āandā, āorā), PRP (pronoun),
and MD (modal auxiliary, e.g., ācanā, āwillā).
Phrases can then be defined as simple noun groups, for
example
23
26. Indexing (Inverted Index)
Recall, indexes are designed to support search.
Each index term is associated with an inverted list
ā¦ Contains lists of documents, or lists of word occurrences in documents,
and other information.
ā¦ Each entry is called a posting.
ā¦ The part of the posting that refers to a specific document or
location is called a pointer
ā¦ Each document in the collection is given a unique number
ā¦ Lists are usually document-ordered (sorted by document number)
26
31. Retrieval Function
Ranking
Documents are retrieved in sorted order according to a score computing using
the document representation, the query, and a ranking algorithm
31
32. Retrieval Function (Boolean Retrieval)
Advantages
ā¦ Results are predictable, relatively easy to explain
ā¦ Many different features can be incorporated
ā¦ Efficient processing since many documents can be eliminated from search
Disadvantages
ā¦ Effectiveness depends entirely on user
ā¦ Simple queries usually donāt work well
ā¦ Complex queries are difficult
32
33. Retrieval Function (Boolean Retrieval)
Sequence of queries driven by number of retrieved documents
ā¦ e.g. ālincolnā search of news articles
ā¦ president AND lincoln
ā¦ president AND lincoln AND NOT (automobile OR car)
ā¦ president AND lincoln AND biography AND life AND birthplace AND gettysburg
AND NOT (automobile OR car)
ā¦ president AND lincoln AND (biography OR life OR birthplace OR gettysburg)
AND NOT (automobile OR car)
33
34. Retrieval Function (Vector Space Model)
Ranked based method.
Documents and query represented by a vector of term weights
Collection represented by a matrix of term weights
34
35. Retrieval Function (Vector Space Model)
borneo daily new north straits times
D1 0 0 1 0 1 1
D2 0 1 1 0 1 0
D3 1 0 0 1 0 1
D1: new straits times
D2: new straits daily
D3 : north borneo times
Vector of terms
35
36. Retrieval Function (Vector Space Model)
borneo daily new north straits times
D1 0 0 0.176 0 0.176 0.176
D2 0 0.477 0.176 0 0.176 0
D3 0.477 0 0 0.477 0 0.176
idf (borneo) = log(3/1) =0.477
idf (daily) = log(3/1) = 0.477
idf (new) = log(3/2) =0.176
idf (north) = log(3/1) = 0.477
idf (straits) = log(3/2) = 0.176
idf (times) = log(3/2) = 0.176
then multiply by tf
tf.idf weight
Term frequency weight measures
importance in document:
Inverse document frequency measures
importance in collection:
36
37. Retrieval Function (Vector Space Model)
Documents ranked by distance between points representing
query and documents
ā¦ Similarity measure more common than a distance or dissimilarity
measure
ā¦ e.g. Cosine correlation
37
38. Retrieval Function (Vector Space Model)
Consider two documents D1, D2 and a query Q
Q = āstraits timesā
Compare against collection,
(borneo, daily, new, north, straits, times)
Q = (0, 0, 0, 0, 0.176, 0.176)
D1 = (0, 0, 0.176, 0, 0.176, 0.176)
D2 = (0, 0.477, 0.176, 0, 0.176, 0)
š¶š¶š¶š¶š¶š¶š¶š¶š¶š¶š¶š¶ š·š·š·, šš =
0ā0 + 0ā0 + 0.176ā0 + 0ā0.176 + 0.176ā0.176 +(0.176ā0.176)
0.1762
+0.1762
+0.1762
(0.1762
+0.1762
)
=0.816
Find Cosine (D2,Q).
Which document is more relevant?
38
39. Evaluation
A must to evaluate the retrieval function, preprocessing steps etc.
Standard Collection
ā¦ Task specific
ā¦ Human experts are used to judge relevant results.
Performance Metric
ā¦ Precision
ā¦ Recall
39
42. Evaluation (Collection)
Obtaining relevance judgments is an expensive, time-consuming
process
ā¦ who does it?
ā¦ what are the instructions?
ā¦ what is the level of agreement?
42
43. Evaluation (Collection)
Exhaustive judgments for all documents in a collection is not
practical
Pooling technique is used in TREC
ā¦ top k results (for TREC, k varied between 50 and 200) from the rankings
obtained by different search engines (or retrieval algorithms) are merged into
a pool
ā¦ duplicates are removed
ā¦ documents are presented in some random order to the relevance judges
Produces a large number of relevance judgments for each query,
although still incomplete
43
47. Search Engine
The most relevant application of
Information Retrieval.
Do you agree?
Search on the Web is a daily activity for
many people throughout the world.
47
48. What about Database Records
Database records (or tuples in relational databases) are typically
made up of well-defined fields (or attributes)
ā¦ e.g., bank records with account numbers, balances, names, addresses, social
security numbers, dates of birth, etc.
Easy to compare fields with well-defined semantics to queries in
order to find matches
Text is more difficult
48
49. Search Query vs DB Query
Example bank database query
ā¦ Find records with balance > $50,000 in branches located in Amherst, MA.
ā¦ Matches easily found by comparison with field values of records
Example search engine query
ā¦ bank scandals in western mass
ā¦ This text must be compared to the text of entire news stories
49
50. Dimensions of IR
IR is more than just text, and more than just web search
ā¦ although these are central
People doing IR work with different media, different types
of search applications, and different tasks
50
51. Other Media
New applications increasingly involve new media
ā¦ e.g., video, photos, music, speech
Like text, content is difficult to describe and compare
ā¦ text may be used to represent them (e.g. tags)
IR approaches to search and evaluation are appropriate
51
52. Dimensions of IR
Content Applications Tasks
Text Web search Ad hoc search
Images Vertical search Filtering
Video Enterprise search Classification
Scanned docs Desktop search Question answering
Audio Forum search
Music P2P search
Literature search
52
53. IR and Search Engines
Relevance
-Effective ranking
Evaluation
-Testing and measuring
Information needs
-User interaction
Performance
-Efficient search and indexing
Incorporating new data
-Coverage and freshness
Scalability
-Growing with data and users
Adaptability
-Tuning for applications
Specific problems
-e.g. Spam
Information Retrieval Search Engines
53
54. Search Engine Issues
Performance
ā¦ Measuring and improving the efficiency of search
ā¦ e.g., reducing response time, increasing query throughput, increasing indexing speed
ā¦ Indexes are data structures designed to improve search efficiency
ā¦ designing and implementing them are major issues for search engines
54
55. Search Engine Issues
Dynamic data
ā¦ The ācollectionā for most real applications is constantly changing in
terms of updates, additions, deletions
ā¦ e.g., web pages
ā¦ Acquiring or ācrawlingā the documents is a major task
ā¦ Typical measures are coverage (how much has been indexed) and freshness (how recently
was it indexed)
ā¦ Updating the indexes while processing queries is also a design
issue
55
56. Search Engine Issues
Scalability
ā¦ Making everything work with millions of users every day, and many
terabytes of documents
ā¦ Distributed processing is essential
Adaptability
ā¦ Changing and tuning search engine components such as ranking
algorithm, indexing strategy, interface for different applications
56
58. Letās Define the Challenge Together
1. Cross Lingual IR
2. Big Data
3. Personalization
4. Domain Specific IR
5. Multi modal IR
6. ā¦.
58
59. IR Research Direction
Latest Research at Google
http://research.google.com/pubs/In
formationRetrievalandtheWeb.html
59
60. Acknowledgement
Thank you for Watching ā¦
This presentation is prepared with
some adapted contents from the
following sources.
a. Introduction to Information
Retrieval, C. D. Manning, P.
Raghavan and H. Schutze, 2009.
b. Introduction to Information
Retrieval, IR Summer School 2001,
Mounia, Lalmas.
c. Search Engines: Information
Retrieval in Practice, D. Metzler, T.
Strohman and W. B. Croft, 2009.
60