Date post: | 17-Jan-2018 |
Category: |
Documents |
Upload: | lee-manning |
View: | 222 times |
Download: | 0 times |
My past worksMy past works
Kazunari Sugiyama16 Apr., 2009
WING Group Meeting
2
OutlineOutlineI. Natural Language Processing (Disambiguation)
I-1 Personal Name Disambiguation in Web Search Results
I-2 Word Sense Disambiguation
in Japanese TextsII. Web Information Retrieval
II-1 Characterizing Web Pages Using Hyperlinked Neighboring Pages
II-2 Adaptive Web Search based on User’s Information Needs
3
I-1 Personal Name Disambiguation I-1 Personal Name Disambiguation in Web Search Resultsin Web Search Results
[Outline]1. Introduction2. Our Proposed Method3. Experiments
4
Politician
Professor of Computer Science
consulting company
Robert M. Gates(other person not “William Cohen”)
5
2.1 2.1 Our Proposed MethodOur Proposed Method
tn
t1
t2
t3
[Semi-supervised Clustering]
×G1
×G2 tn
t1
t2
t3
×G1
×
×G’1D(G, w )p
Control the fluctuation of the centroid of a cluster
G2
p : search-result Web page
w p: feature vector of pGi : the centroid vector of a cluster
psi: seed page
wp: feature vector of a Web page contained in a cluster that has centroid G
(G)
w pw ps1
w ps2
w p
w ps1w ps2
21
),()(
)(
nnD p
pn p
newp
wGww
GGw
G
G
C1C2 C1
C2
6
3. Experiments3. Experiments
3.1 Experimental Data3.2 Evaluation Measure3.3 Experimental Results
7
3.1 3.1 Experimental DataExperimental Data WePS Corpus
– Established for “Web People Search Task” at SemEval-2007 in the Association for Computational Linguistics (ACL) conference
Web pages related to 79 personal names– Sampled from
• participants in conferences on digital libraries and computational linguistics,
• bibliographic articles in the English Wikipedia,• the U.S. Census
– The top 100 Yahoo search results via its search API for a personal name query• Training sets : 49 names, Test sets : 30 names ( 7,900 Web pages
in total )
Pre-processing for WePS corpus– Eliminate stopwords, and perform stemming– Determine the optimal parameter for merging similar
clusters using the training set in the WePS corpus and then apply it to the test set in the WePS corpus
8
3.2 3.2 Evaluation MeasureEvaluation Measure
(1) Purity(2) Inverse purity(3) F (harmonic mean of (1) and (2))
(These are standard evaluation measures employed in the “Web People search task.”)
[Hotho et al., GLDV Journal’05]
9
3.3 3.3 Experimental ResultsExperimental Results
Team ID of “Web People Search Task”
Purity Inverse Purity F
CU_COMSEM 0.72 0.88 0.78IRST-BP 0.75 0.80 0.75PSNUS 0.73 0.82 0.75UVA 0.81 0.60 0.67SHEF 0.60 0.82 0.66Our proposed method 2 and 3 sentences in 5 Wikipedia seed pages and search result Web page, redpectively
0.72 0.81 0.76
This result is comparable to the top results (0.78) among the top 5 participants in “Web People Search Task.”
We could acquire useful information from sentences that characterize an entity of a person and disambiguate the entity of a person effectively.
10
I-2. Word Sense Disambiguation I-2. Word Sense Disambiguation in Japanese Textsin Japanese Texts
[Outline]1. Introduction2. Proposed Method3. Experiments
11
1. 1. Introduction (1/2)Introduction (1/2)
Word Sense Disambiguation (WSD)– Determining the meaning of ambiguous word
in its context“run”:
(1) Bob goes running every morning.“to move fast by using one’s feet”
(2) Mary runs a beauty parlor.“to direct or control”
12
1. 1. Introduction (2/2)Introduction (2/2)
Raw corpusnot assigned sense tags
Add sense-tagged instances
(2)
(3)
Feature extractionfor clustering and WSD(“baseline features”)
(1)
Semi-supervised clustering
SupervisedWSD
Feature extractionfor WSD from clustering results
(4)
(“seed instances”)
Our approach for WSD Basically, supervised WSD Applying semi-supervised clustering
by introducing sense-tagged instances to supervised WSD
13
2. 2. Proposed MethodProposed Method
2.1 Semi-supervised Clustering2.2 Features for WSD
obtained Using Clustering Results
14
2.1 Semi-supervised Clustering2.1 Semi-supervised Clustering
2.1.1 Features for Clustering2.1.2 Semi-supervised Clustering2.1.3 Seed Instances and Constraints
for Clustering
15
2.1.1 2.1.1 Features for Clustering and WSD Features for Clustering and WSD (“baseline features”)(“baseline features”)
Morphological features– Bag-of-words (BOW), Part-of-speech (POS),
and detailed POS classification Target word itself and the two words to its right and left.
Syntactic features– If the POS of a target word is a noun, extract the verb in a
grammatical dependency relation with the noun.– If the POS of a target word is a verb, extract the noun in a
grammatical dependency relation with the verb. Figures in Bunrui-Goi-Hyou
– 4 and 5 digits regarding the content word to the right and left of the target word.
地域 (“community”) 社会 (“society”) “1.1720,4,1,3” → 1172 (as 4 digits), 11720 (as 5digits)
5 topics inferred on the basis of LDA– Compute the log-likelihood of a word instance
(“soft-tag” approach [Cai et al., EMNLP’07])
16
2.1.2 Semi-supervised Clustering2.1.2 Semi-supervised Clustering[Proposed method]
x : word instance
Gi: the centroid of a clusterx si
: seed instancecD
nn
cD
CC
CC
xn
C
xn
Cnew
xx
),(1
),(1
21
2121
21 GG
GGff
Gff
tn
t1
t2
t3
×G1
×G2 w ps2C1
C3
× G2C2
xs1
tn
t1
t2
t3
×G1
×G2 w ps2C1
C3
× G2C2
xs1
Gnew×
Cnew
),( 21 CCD GG : adaptive Mahalanobis distance
Refine [Sugiyama and Okumura, ICADL’07] (about Web People Search Task)for word instances
Control the fluctuation of the centroid of a cluster
17
2.1.3 Seed Instances 2.1.3 Seed Instances and Constraints for Clustering (1/3)and Constraints for Clustering (1/3)
Select initial seed instances:(I-1) randomly,(I-2) “KKZ”(I-3) centroid of a cluster
generated by K-means initial instances for K-means * randomly,
* “KKZ” [Katsavounidis et al., IEEE Signal Processing Letters, ‘94]
[Method I]Data set of word instances
Training data set Data set for clustering
: seed instance
18
2.1.3 Seed Instances 2.1.3 Seed Instances and Constraints for Clustering (2/3)and Constraints for Clustering (2/3)
[Method II]Data set of word instances
Training data set
s1, s2, s3: word senses of target word
s1 s2 s3
Data set for clustering
: seed instance
Select initial seed instances:(II-1) By considering the frequency
of word senses(II-1-1) randomly,(II-1-2) “KKZ”(II-1-3) centroid of a cluster generated by K-means
initial instances for K-means * randomly, * “KKZ”
(II-2) In proportion to the frequency of word senses (D’Hondt
method)(II-2-1) randomly,(II-2-2) “KKZ”(II-2-3) centroid of a cluster generated by K-means initial instances for K-means
* randomly, * “KKZ”
19
Constraints– “cannot-link” only,– “must-link” only,– Both constraints, – “cannot-link” and “must-link” without outliers
2.1.3 Seed Instances 2.1.3 Seed Instances and Constraints for Clustering (3/3)and Constraints for Clustering (3/3)
Put “must-link” constraints between and
×
×is an outlier,
so a “must-link” constraint is not added.
Csw
Csv
Csv
Csw
C sv C sw
C sw
D( , ) < , G new G sC v ThdisD( , ) < G new G sC w ThdisG new
G newD( , ) > , G new G sC v ThdisD( , ) > G new G sC w Thdis
D( , ) Gnew svGCD( , ) Gnew swGC
D( , ) Gnew swGC
D( , ) Gnew svGC
(Thdis = 0.388)
[“must-link” without outliers]
20
2.2 2.2 Features for WSD Obtained Using Features for WSD Obtained Using Clustering ResultsClustering Results
(a) Inter-cluster information TF in a cluster (TF) Cluster ID (CID) Sense frequency (SF)
(b) Context information regarding adjacent words
Mutual information (MI) T-score (T) (CHI2)
(c) Context information regarding two words to the right and left of the target word
Information gain (IG)
2
)1,...,2(1 iww ii
21012 wwwww
*We employ (b) and (c) to reflect the concept of “one sense per collocation.”[Yarowsky, ACL’95]
21
4. 4. ExperimentsExperiments
4.1 Experimental Data4.2 Semi-supervised Clustering4.3 Word Sense Disambiguation
22
4.1 4.1 Experimental DataExperimental Data
RWC corpus from the “SENSEVAL-2 Japanese Dictionary Task”– 3000 Japanese newspaper articles issued in
1994 Sense tags in Japanese Dictionary, “Iwanami
Kokugo Jiten” were manually assigned to 148,558 ambiguous words.
– 100 target words 50 nouns, and 50 verbs
23
4.2 4.2 Semi-supervised ClusteringSemi-supervised Clustering
Our semi-supervised clustering approach outperforms other distance-based approaches.– Our method locally adjusts the centroid of a cluster
Comparison of distance-based approaches
[Observations]
Distance-based approaches
24
4.3 4.3 Word Sense DisambiguationWord Sense Disambiguation Experimental Results
[Observations]• The best accuracy is obtained, when we add features from clustering results, CID, MI, and IG to the baseline features.
• According to the results of OURS, TITECH, NAIST, WSD accuracy issignificantly improved by adding features computed from clustering results.
25
II-1 Characterizing Web Pages Using II-1 Characterizing Web Pages Using Hyperlinked Neighboring PagesHyperlinked Neighboring Pages
[Outline]1. Introduction2. Our Proposed Method3. Experiments
26
1. Introduction1. Introduction
Transitions of Web Search EnginesThe first-generation search engines:
• Only terms included in Web pages were utilized as indices of Web pages.• Peculiar features of the Web such as hyperlink structures are not exploited.
The second-generation search engines: • The hyperlink structures of Web pages are considered.
e.g., (1) “Optimal Document Granularity” based IR systems(2) HITS (CLEVER project), Teoma (DiscoWeb)(3) PageRank (Google)
Users are not satisfied with• the ease of use,• retrieval accuracy.
27
2. Our Proposed Methods2. Our Proposed Methods
t1
tm
t2
t3
feature vector of :
ptgttarget Web page:
ptgt
feature vector of Web page hyperlinked from ptgt
w ptgt
pw’ tgt
w ptgt
refined feature vectorof : ptgt
pw’ tgt
28
Method Method ⅠⅠ
ptgt
[in the Web space]
tm
t2
t3
[in the vector space]
t1
w ptgt pw’ tgtL(in)
L(out)
29
Method ⅡMethod Ⅱ
ptgt
tm
t2
t3
t1clusters generated from groups of Web pagesin each level from target page
w ptgt pw’ tgt
centroid vectorof clusters
[in the Web space] [in the vector space]
L(in)
L(out)
30
Method Method ⅢⅢ
ptgt
tm
t2
t3
t1clusters generated from groups of Web pagesin each level from target page
w ptgt pw’ tgt
centroid vectorof clusters
[in the Web space] [in the vector space]
L(in)
L(out)
31
Experimental SetupExperimental Setup
Data Set– TREC WT10g test collection
(10GB, 1.69 million Web pages)
Specification of Workstation– Sun Enterprise 450
CPU: UltraSparc-II 480MHz Memory: 2 GB OS: Solaris 8
32
Web pages Using in the ExperimentWeb pages Using in the Experiment
ptgt
33
Experimental ResultsExperimental ResultsComparison of best search accuracy
obtained using each Method I, II, III
Comparison of best search accuracyobtained using each Method I, II, III
00.10.20.30.40.50.60.7
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1recall
prec
ision TF- IDF
(MI-a), L(in)=3(MII-a), L(in)=1, K=2(MIII-a), L(in)=2, K=3
The contents of a target Web page can be represented much better by making a point of in-linked pages of a target page.
tfidf(MI-a), L(in)=3
averageprecision
% improvement
0.3130.3420.3400.345
+2.9+2.7+3.2
-
(MII-a), L(in)=1, K=2
(MIII-a), L(in)=2, K=3
(MI-a), L(in)=3
averageprecision % improvement
0.032
0.3400.345
+31.0+30.8+31.3
(MII-a), L(in)=1, K=2
(MIII-a), L(in)=2, K=3
HITS
modified HITS
modified HITS with weighted links
0.136
0.1380.342
-+10.4
+10.6
34
Method Method ⅢⅢ
ptgt
tm
t2
t3
t1clusters generated from groups of Web pagesin each level from target page
w ptgt pw’ tgt
centroid vectorof clusters
[in the Web space] [in the vector space]
L(in)
1
2
35
II-2 Adaptive Web Search Based on II-2 Adaptive Web Search Based on User’s Information NeedsUser’s Information Needs
[Outline]1. Introduction2. Our Proposed Method3. Experiments
36
1. Introduction1. Introduction The World Wide Web (WWW)
– It has become increasingly difficult for users to find relevant information.
– Web search engines help users find useful information on the WWW.
Web Search Engines– Return the same results regardless of who submits the query.
Web search results should adapt to users with different information
needs.
In general, each user has different information needs for his/her query.
37
2. Our Proposed Method2. Our Proposed Method
User profile construction based on(1) pure browsing history,(2) modified collaborative filtering.
38
(1) User Profile Construction Based (1) User Profile Construction Based on Pure Browsing History (1/3)on Pure Browsing History (1/3)
We assume that the preferences of each user consist of the following two aspects:
Pand construct user profile .
P per(1) Persistent (or long term) preferences ,
P today(2) Ephemeral (or short term) preferences
39
(1) User Profile Construction Based (1) User Profile Construction Based on Pure Browsing History (2/3)on Pure Browsing History (2/3)
Browsing history of today(0 day ago)
Browsing history of 2 days ago
1 SN
Browsing history ofN days ago
[Persistent preferences]
: Web page : Window size
[Ephemeral preferences]
1 S2
Browsing history of 1 days ago
1 S1
1 S0
1st browsing history in today
(1) 1 S0
browsing history in today
(n )bh
nbh th
1 hp
rth browsing history in today
(r) S0(r)(1) (r) 1 S0
(cur)(cur)
current sessionP (r)
P (br)P(cur)
Pper
P today
40
(1) User Profile Construction Based (1) User Profile Construction Based on Pure Browsing History (3/3)on Pure Browsing History (3/3)
User profile is finally constructed as follows: P
).5.0,5.0(,1satisfythat constantsareandand,1satisfythat constantsareandwhere
)()(
yxyxyxbaba
bybxa
bacurbrper
todaypertoday
PPPPPP (a)
(b)
41
Overview of the Pure Overview of the Pure Collaborative Filtering AlgorithmCollaborative Filtering Algorithm
user 1user 2
user a
user U
item 1 item 2 item i item I
Item that prediction is computed
Active user
25
3 21 4
5
5
3
24
User-item ratings matrix for collaborative filtering
42
(2) User Profile Construction Based on (2) User Profile Construction Based on Modified Collaborative FilteringModified Collaborative Filtering
(( 1/21/2 ))User-term weight matrix for modified collaborative filtering
user 1
user 2
user a
user U
term 1 term 2 term i term TTerm weight that prediction is computed
Active user
user 1
user 2
user a
user U
term 1 term 2 term i term T term T+1 term T+2 term T+v
0.745
0.745 0.362
0.835
0.639
0.4610.247
0.534
0.928
0.718
0.126
0.485
0.718
0.126
0.485
0.534
0.928
0.362
0.835
0.639
0.4610.247
0.723
0.686
0.328
0.451
0.563
0.172
When each user browsed k Web pages
Active user
Term weight that prediction is computed
When each user browsed k+1 Web pages
43
User Profile Construction User Profile Construction Based on Based on
Modified Collaborative Filtering AlgorithmModified Collaborative Filtering Algorithm
User Profile Construction Based on (1) Static Number of Users in the Neighborhood,
(2) Dynamic Number of Users in the Neighborhood.
44
ExperimentExperiment (1/2)(1/2)
Explicit Method(1) Relevance Feedback ,
Implicit Method(2) Method based on pure browsing history ,
(3) Method based on modified collaborative filtering
Construction of User Profile
Observed browsing history– 20 subjects – 30 days
45
Experiments (2/2)Experiments (2/2)Query
– 50 query topics employed as test topics in the TREC WT10g test collection
Evaluation– Compute similarity between user profile and
feature vector of Web page in search results– R-precision (R=30)
46
User Profile Based on User Profile Based on Pure Browsing History (Implicit Method)Pure Browsing History (Implicit Method)
Using , and , user profile is defined as follows:
P per P (br)P
(cur) P
1where
)()(
babybxa curbrper PPPP
47
Experimental ResultsExperimental Results
• A user profile that provides search results adaptive to a user can be constructed when a window size with about 15 days is used.
• This approach can achieve about 3% higher precision than the relevance feedback-based user profile.
The user’s browsing history strongly reflectsthe user’s preference.
48
User Profile Based on User Profile Based on Modified Collaborative Filtering Modified Collaborative Filtering
(Implicit Method)(Implicit Method)Using , and , user profile is defined as follows:
Pper P (br) V (pre)
P
1where
)()(
babybxa prebrper VPPP
User Profile Construction Based on (1) Static Number of Users in the Neighborhood,(2) Dynamic Number of Users in the Neighborhood
49
Experimental ResultsExperimental Results(Dynamic Method)(Dynamic Method)
• In all of our experimental results, the best precision is obtained in the case of x=0.129, y= 0.871.
In this method, the neighborhood of each user is determined by the centroid vectors of clusters of users, and the number of clusters is different user by user.
This allows each user to perform more fine-grained search compared with the static method.
50
Thank you very much!Thank you very much!