Date post: | 18-Jan-2016 |
Category: |
Documents |
Upload: | phyllis-douglas |
View: | 219 times |
Download: | 1 times |
Text Operations
J. H. WangFeb. 21, 2008
The Retrieval ProcessUserInterface
Text Operations
Query Operations
Indexing
Searching
Ranking
Index
Text
query
user need
user feedback
ranked docs
retrieved docs
logical viewlogical view
inverted file
DB Manager Module
4, 10
6, 7
5 8
2
8
Text Database
Text
Outline
• Document Preprocessing (7.1-7.2)• Text Compression (7.4-7.5): skipped• Automatic Indexing (Chap. 9, Salton)
– Term Selection
Document Preprocessing
• Lexical analysis– Letters, digits, punctuation marks, …
• Stopword removal– “the”, “of”, …
• Stemming– Prefix, suffix
• Index term selection– Noun
• Construction of term categorization structure– Thesaurus
• Logical view of the documents
structure
Accents,spacing stopwords
Noungroups stemming
Manual indexingDocs
structure Full text Index terms
Lexical Analysis
• Converting a stream of characters into a stream of words– Recognition of words– Digits: usually not good index terms
• Ex.: The number of deaths due to car accidents between 1910 and 1989, “510B.C.”, credit card numbers, …
– Hyphens• Ex: state-of-the-art, gilt-edge, B-49, …
– Punctuation marks: normally removed entirely• Ex: 510B.C., program codes: x.id vs. xid, …
– The case of letters: usually not important• Ex: Bank vs. bank, Unix-like operating systems, …
Elimination of Stopwords• Stopwords: words which are too freque
nt among the documents in the collection are not good discriminators– Articles, prepositions, conjunctions, …– Some verbs, adverbs, and adjectives
• To reduce the size of the indexing structure
• Stopword removal might reduce recall– Ex: “to be or not to be”
Stemming
• The substitution of the words by their respective stems– Ex: plurals, gerund forms, past tense suffixes, …
• A stem is the portion of a word which is left after the removal of its affixes (i.e., prefixes and suffixes)– Ex: connect, connected, connecting,
connection, connections
• Controversy about the benefits– Useful for improving retrieval performance– Reducing the size of the indexing structure
Stemming
• Four types of stemming strategies– Affix removal, table lookup, successor
variety, and n-grams (or term clustering)
• Suffix removal– Port’s algorithm (available in the
Appendix)• Simplicity and elegance
Index Term Selection
• Manually or automatically• Identification of noun groups
– Most of the semantics is carried by the noun words
– Systematic elimination of verbs, adjectives, adverbs, connectives, articles, and pronouns
– A noun group is a set of nouns whose syntactic distance in the text does not exceed a predefined threshold
Thesauri
• Thesaurus: a reference to a treasury of words– A precompiled list of important words in
a given domain of knowledge– For each word in this list, a set of related
words• Ex: synonyms, …
– It also involves normalization of vocabulary, and a structure
Example Entry in Peter Roget’s Thesaurus
• Cowardly adjective• Ignobly lacking in courage: cowardly tur
ncoats.• Syns: chicken (slang), chicken-hearted,
craven, dastardly, faint-hearted, gutless, lily-livered, pusillanimous, unmanly, yellow (slang), yellow-bellied (slang).
Main Purposes of a Thesaurus
• To provide a standard vocabulary for indexing and searching
• To assist users with locating terms for proper query formulation
• To provide classified hierarchies that allow the broadening and narrowing of the current query request according to the user needs
Motivation for Building a Thesaurus
• Using a controlled vocabulary for the indexing and searching– Normalization of indexing concepts– Reduction of noise– Identification of indexing terms with a
clear semantic meaning– Retrieval based on concepts rather than
on words• Ex: term classification hierarchy in
Yahoo!
Main Components of a Thesaurus
• Index terms: individual words, group of words, phrases– Concept
• Ex: “missiles, ballistic”
– Definition or explanation • Ex: seal (marine animals), seal (documents)
• Relationships among the terms– BT (broader), NT (narrower)– RT (related): much difficult
• A layout design for these term relationships– A list or bi-dimensional display
Automatic Indexing(Term Selection)
Automatic Indexing
• Indexing– assign identifiers (index terms) to text
documents
• Identifiers– single-term vs. term phrase– controlled vs. uncontrolled vocabularies
instruction manuals, terminological schedules, …
– objective vs. nonobjective text identifiers cataloging rules control, e.g., author names, publisher names, dates of publications, …
Two Issues
• Issue 1: indexing exhaustivity– exhaustive: assign a large number of terms– Nonexhaustive: only main aspects of subject conte
nt• Issue 2: term specificity
– broad terms (generic)cannot distinguish relevant from nonrelevant documents
– narrow terms (specific)retrieve relatively fewer documents, but most of them are relevant
All docs
Recall vs. Precision
• Recall (R) = Number of relevant documents retrieved / total number of relevant documents in collection– The proportion of relevant items retrieved
• Precision (P) = Number of relevant documents retrieved / total number of documents retrieved– The proportion of items retrieved that are
relevant
• Example: for a query, e.g. TaipeiRetrieveddocs
Relevantdocs
More on Recall/Precision
• Simultaneously optimizing both recall and precision is not normally achievable– Narrow and specific terms: precision is favored– Broad and nonspecific terms: recall is favored
• When a choice must be made between term specificity and term breadth, the former is generally preferable– High-recall, low-precision documents will
burden the user– Lack of precision is more easily remedied than
lack of recall
Term-Frequency Consideration
• Function words– for example, "and", "of", "or", "but", …– the frequencies of these words are high in
all texts• Content words
– words that actually relate to document content
– varying frequencies in the different texts of a collection
– indicate term importance for content
A Frequency-Based Indexing Method
• Eliminate common function words from the document texts by consulting a special dictionary, or stop list, containing a list of high frequency function words
• Compute the term frequency tfij for all remaining terms Tj in each document Di, specifying the number of occurrences of Tj in Di
• Choose a threshold frequency T, and assign to each document Di all term Tj for which tfij > T
More on Term Frequency
• High-frequency term– Recall
• Ex: “Apple”
– But only if its occurrence frequency is not equally high in other documents
• Low-frequency term– Precision
• Ex: “Huntington’s disease”
– Able to distinguish the few documents in which they occur from the many from which they are absent
How to Compute Weight wij ?• Inverse document frequency, idfj
– tfij*idfj (TFxIDF)
• Term discrimination value, dvj
– tfij*dvj
• Probabilistic term weighting trj
– tfij*trj
• Global properties of terms in a document collection
Inverse Document Frequency
• Inverse Document Frequency (IDF) for term Tj
where dfj (document frequency of term Tj) is thenumber of documents in which Tj occurs.
– fulfil both the recall and the precision– occur frequently in individual documents but
rarely in the remainder of the collection
idfN
dfj
j
log
TFxIDF• Weight wij of a term Tj in a document di
• Eliminating common function words• Computing the value of wij for each term Tj in e
ach document Di
• Assigning to the documents of a collection all terms with sufficiently high (tf x idf) weights
w tfN
dfij ij
j
log
Term-discrimination Value
• Useful index terms– Distinguish the documents of a collection from
each other
• Document Space– Each point represents a particular document of
a collection– The distance between two points is inversely
proportional to the similarity between the respective term assignments
• When two documents are assigned very similar term sets, the corresponding points in document configuration appear close together
Original State After Assignment of good discriminator
After Assignment of poor discriminator
A Virtual Document Space
Good Term Assignment
• When a term is assigned to the documents of a collection, the few documents to which the term is assigned will be distinguished from the rest of the collection
• This should increase the average distance between the documents in the collection and hence produce a document space less dense than before
Poor Term Assignment
• A high frequency term is assigned that does not discriminate between the objects of a collection
• Its assignment will render the document more similar
• This is reflected in an increase in document space density
Term Discrimination Value
• Definitiondvj = Q - Qj
where Q and Qj are space densities before and after the assignment of term Tj
• The average pairwise similarity between all pairs of distinct terms:
• dvj>0, Tj is a good term; dvj<0, Tj is a poor term.
QN N
sim D Di kki k
N
i
N
1
1 11( )( , )
DocumentFrequency
Low frequency
dvj=0Medium frequency
dvj>0
High frequency
dvj<0
N
Variations of Term-Discrimination Valuewith Document Frequency
TFij x dvj
• wij = tfij x dvj• compared with
– : decreases steadily with increasing document frequency
– dvj: increases from zero to positive as the document frequency of the term increases,
decreases shapely as the document frequency becomes still larger
• Issue: efficiency problem to compute N(N-1) pairwise similarities
w tfN
dfij ij
j
log
N
df j
Document Centroid• Document centroid C = (c1, c2, c3, ..., c
t)
where wij is the j-th term in document I– A “dummy” average document located in
the center of the document space• Space density
N
iijj wc
1
N
iiDCsim
NQ
1
),(1
Probabilistic Term Weighting
• GoalExplicit distinctions between occurrences of terms in relevant and nonrelevant documents of a collection
• DefinitionGiven a user query q, and the ideal answer set of the relevant documentsFrom decision theory, the best ranking algorithm for a document D
)Pr(
)Pr(log
)|Pr(
)|Pr(log)(
nonrel
rel
nonrelD
relDDg
Probabilistic Term Weighting
• Pr(rel), Pr(nonrel):document’s a priori probabilities of relevance and nonrelevance
• Pr(D|rel), Pr(D|nonrel):occurrence probabilities of document D in the relevant and nonrelevant document sets
t
ii
t
ii
nonrelxnonrelD
relxrelD
1
1
)|Pr()|Pr(
)|Pr()|Pr(
Assumptions
• Terms occur independently in documents
Derivation Process
)Pr(
)Pr(log
)|Pr(
)|Pr(log)(
nonrel
rel
nonrelD
relDDg
log
Pr( | )
Pr( | )
x rel
x nonrel
ii
t
ii
t1
1
constants
log
Pr( | )
Pr( | )
x rel
x nonreli
ii
t
1
constants
• Given a document D=(d1, d2, …, dt)
• Assume di is either 0 (absent) or 1 (present)
Pr( | ) ( )
Pr( | ) ( )
x d rel p p
x d nonrel q q
i i i
d
i
d
i i i
d
i
d
i i
i i
1
1
1
1
Pr(xi=1|rel) = pi Pr(xi=0|rel) = 1-piPr(xi=1|nonrel) = qi Pr(xi=0|nonrel) = 1-qi
g Dx d rel
x d nonreli i
i ii
t
( ) logPr( | )
Pr( | )
1
constants
For a specific document D
g Dx d rel
x d nonreli i
i ii
t
( ) logPr( | )
Pr( | )
1
constants
log( )
( )
d d
d d
i i
i i
p p
q q
i i
i ii
t1
11
11
constants
log( ) ( )
( ) ( )
d d
d d
i i
i i
p q p
q p qi i i
i i ii
t 1 1
1 11
constants
constantslog1 )1())1((
)1())1((
t
iiii
iii
qpq
pqpi
i
d
d
trp q
q pj
j j
j j
log( )
( )
1
1
g Dp
qd
p q
q pi
ii
t
ii i
i ii
t
( ) log log( )
( )
1
1
1
11 1constants
Term Relevance Weight
Issue
• How to compute pj and qj ?
pj = rj / Rqj = (dfj-rj)/(N-R)
– rj: the number of relevant documents that contains term Tj
– R: the total number of relevant documents– N: the total number of documents
Estimation of Term-Relevance
• The occurrence probability of a term in the nonrelevant documents qj is approximated by the occurrence probability of the term in the entire document collection
qj = dfj / N– Large majority of documents will be nonrelevant to
the average query• The occurrence probabilities of the terms in th
e small number of relevant documents is assumed to be equal by using a constant value pj = 0.5 for all j
5.0*
)1(*5.0log
)1(
)1(log
N
dfN
df
pq
qptr
j
j
jj
jjj
j
j
df
dfN )(log
When N is sufficiently large, N-dfj N,
j
jj
df
dfNtr
)(log
jdf
Nlog = idfj
Comparison