+ All Categories
Home > Documents > Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive...

Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive...

Date post: 25-Aug-2020
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
47
Text Mining Using Linear Models of Latent States Bob Stine Department of Statistics The Wharton School, University of Pennsylvania www-stat.wharton.upenn.edu/~stine
Transcript
Page 1: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Text Mining Using Linear Models

of Latent States

Bob StineDepartment of Statistics

The Wharton School, University of Pennsylvaniawww-stat.wharton.upenn.edu/~stine

Page 2: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

TopicsApplication

Statistical named entity recognition

Feature creationPreprocessingConverting text into numerical data

Exploiting the featuresEstimators, standard errorsAuctions and experts

CollaboratorsDean Foster in StatisticsLyle Ungar in CS

2

Page 3: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Application andMotivation

Page 4: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Text Mining ApplicationsCloze

What’s the next word?“...in the midst of modern life the greatest, __”Data compression

Word disambiguationMeaning of a word in contextDoes “Washington” refer to a state, a person, a city or perhaps a baseball team? Or politics?

Speech taggingIdentifying parts of speechDistinguishing among proper nouns

Grading papers, classification, ...

4

Page 5: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Named Entity RecognitionAnnotate plain text in a way that identifies the words that refer to a ! ! person (Obama)! ! place (France)! ! organization (Google)or something else.Wiki exampleJim bought 300 shares of Acme Corp in 2006.

Customized systems build on grammatical heuristics and statistical models.

Time consuming to buildSpecific to training domain

5

person company year

Page 6: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Second ExampleYou get some text, a sequence of “words”

bob went to the 7-11 <.> he was hungry <.> ...

Task is to tag proper nouns, distinguishing those associated with people, places and organizations.No other information in the test setTraining data

Marked up sequence that includes the tags that you’d ideally producebob went to the 7-11 <.> he was hungry <.> …

Test data is just a sequence of “words”

6

person organization

Washington?personplaceteam

politics

Page 7: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

ApproachesNumerous methods used for NER

Gazettelists of proper words/businesses, places

Formal grammar, parse treesoff the shelf parsing of text into subject/verb

Stemming such as noting prior word ends in -ing

Capitalization

Not using any of these…Things like capitalization are not available in some formats, such as text from speechGeneralization: gazettes depend on contextLanguages other than English

7Could add these later!

Page 8: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Statistical Models for TextMarkov chains

Hidden Markov models have been successfully used in text mining, particularly speech tagging

Hidden Markov model (HMM)Transition probabilities for observed words! P(wt|wt-1,wt-2,…) as in !P(clear|is,sky,the)Instead specify model for underlying types! P(Tt|Tt-1,Tt-2, …) as in P(adj|is,noun,article)with words generated by the state

8

Tt-2 TtTt-1

wt-1wt-2 wt

Concentrate dependence in

transitions among relatively few

states

Page 9: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

State-Based ModelAppealing heuristic of HMMMeaning of text can be described by transitions in a low-dimensional subspace determined by surrounding textEstimation of HMM hard and slow

NonlinearIterative (dynamic programming)

ObjectiveLinear method for building features that represent underlying state of the text process.Possible? Observable operator algebras for HMMs.

Features used by predictive model. Pick favorite.

9

Page 10: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

ConnectionsTalks earlier today…

Probabilistic latent semantic analysisNon-negative matrix factorization (NMF)Clustering

10

Page 11: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Building the Features

Page 12: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Summary of MethodAccumulate correlations between word occurrences in n-grams

Preprocessing, all n-grams on InternetTrigrams in example; can use/combine with others

Perform a canonical correlation analysis (CCA) of these correlations

Singular value decomposition (SVD) of corr mat

Coordinates of words in the space of canonical variables define “attribute dictionary”Predictive features are sequences of these coordinates determined by the order of the works in the text to be modeled

12

Page 13: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Canonical CorrelationCCA mixes linear regression and principal components analysisRegressionFind linear combination of X1,…,Xk most correlated with Y! max corr(Y, β1X1 + β2X2 + … + βkXk)Canonical correlationFind linear combinations of X’s and Y’s that have maximal correlation! max corr(α1Y1+…+αjYj, β1X1+…+βkXk)Solution is equivalent to PCA of! ! ! ! (ΣXX)-1/2 ΣXY (ΣYY)-1/2

13covariance matrices

Page 14: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Coincidence Matrices

14

billions of n-grams

Pre-word Word Post-wordw1,w2,w3,…,wd w1,w2,w3,…,wd w1,w2,w3,…,wd

0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0

B1 B2Bw

d = 50,000 d is the size of our dictionary

w1,w2,w3

wt-1,wt,wt+1

wn-2,wn-1,wn

...

...

Page 15: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Using CCAWhich words, or groups of words, co-occur?LinearFind α1 in Rd and β1 in R2d that together! ! ! maximize corr(Bwα, [B1,B2]β)(α1,β1) defines first pair of canonical variablesSubsequent pairs as in principle componentsFind (α2,β2) which! ! ! maximize corr(Bwα, [B1,B2]β)while being orthogonal to (α1,β1).We compute about K=30 to 100 of these canonical coordinates

15

Page 16: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Canonical VariablesSVD of correlations C ≈ Bw’[B1 B2]! ! ! C = U D V’ = UD[V1’ V2’]

Attribute dictionary

16

(50,000 x 50)

UD V1 V2

w1

w2

w50000

Wordsin

dict

K=50 columns in each bundle

(50 x 50) (50 x 100,000)

Page 17: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Random ProjectionsFaster calculation of CCA/SVDComputing canonical variables! ! ! ! C = Bw’[B1 B2]50,000 x 100,000 is largeRandom projection

Low rank approximationsReference Halko, Martinsson, Tropp 2010Two stage approach(1) Project into “active” subspace(2) Do usual operation

17

Page 18: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Algorithm for SVDWant SVD of correlations (omit scaling)! ! ! C = Bw’[B1 B2] = UDV’Find orthonormal Q with K+m columns for which! ! ! ! ||C - QQ’C||2 is smallRandom projection! ! Q~N(0,1) works very well!Steps

Compute coefficients H = Q’CSVD of H is U1DV’Compute U = QU1

To get rank K, need a few extra columns (m)

18

Page 19: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Plots of Attribute DictIsolate the coordinates in the attribute dictionary assigned to “interesting words”

Words were not picked out in advance or known while building the attribute dictionary

Several viewsGrouped/colored by parts of speechNamesCommon US names, casual and formalBob and Robert

Numbers

Plots show projections of the coordinates in the attribute dictionary...

19

Page 20: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Parts of SpeechProjection from attribute dictionary

20

nounverbadjunk

Words from d=10,000 dictionary

Not in dictionary

Page 21: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Closer Look at FeaturesFocus on a few names

21

-0.2 -0.1 0.0 0.1 0.2 0.3 0.4

-0.2

-0.1

0.0

0.1

0.2

0.3

0.4

PC 2

PC

3

johndavidmichael paul

robert

george

thomas

william

maryrichard

miketom

charles

bobjoe

joseph

daniel

dan

elizabeth

jennifer

barbarasusan christopher

lisa

linda maria

donaldnancy

karen

margaret

helen

patricia

bettyliz

dorothybetsy

barb

susie

tricia

<OOV>

<s>

</s> ,

the

.

<unk>

-

of

and

:to

a ) (in

|for

QUOTE is/

on

thatby

this with

i's

youit &not or1beare

[]

from at

$

>

as

!

your...

all ;2

0

?

havenewmore an'

was

we

will3

home

canus

about

%if=

2005

page

my

has4

search

free

*

but

our

one

otherdono information5

time

+

they

site

he

up

may

what

which

their

--

news

out

use#any

there

johndavidmichael paul

robert

george

thomaswilliam

maryrichard

miketom

charles

bobjoe

joseph

daniel

dan

elizabeth

jennifer

barbarasusan christopher

lisa

linda mariadonald

nancykaren

margaret

helen

patricia

bettyliz

dorothybetsy

barb

susie

tricia

-0.2 -0.1 0.0 0.1

-0.10

-0.05

0.00

0.05

0.10

PC 2

PC

3

john

davidmichaelpaul

robert

george

thomas

william

maryrichard

miketom

charles

bobjoe

joseph

daniel

dan

elizabeth

jennifer

barbarasusan

christopher

lisa

linda maria

donald

nancy

karen

margaret

helen

patricia

bettyliz

dorothy

betsy

barb

susie

tricia

<s>

<unk> i

0

we%

2005

but

they

he which

there

john

davidmichaelpaul

robert

george

thomas

william

maryrichard

miketom

charles

bobjoe

joseph

daniel

dan

elizabeth

jennifer

barbarasusan

christopher

lisa

linda maria

donald

nancy

karen

margaret

helen

patricia

bettyliz

dorothy

betsy

barb

susie

tricia

Page 22: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Closer Look at FeaturesNumbers as words and digits

22

-0.3 -0.2 -0.1 0.0 0.1 0.2 0.3

-0.1

0.0

0.1

0.2

0.3

PC 2

PC

3

1 234

one

510678

9

two

three

fourfivesixten

seveneight

nine

<OOV><s>

</s>

, the

.

<unk>-

of

and:

to a

)

(

in

|

for

QUOTE

is

/

onthat

bythis

with

i

's

you it

&

not

or

1

be

are

[]

from at

$

>

as!

your... all;

2

0

?

have

new

more

an

'

waswewill

3home

can

us

about

%

if

=

2005

page

myhas

4

search

free*

but

ourone

other

do

no

information

5time

+

they

site

he

up

may

what

whichtheir--

news

out

use

#

any

there

1 234

one

5106789

twothreefourfivesixten

seveneight

nine

Page 23: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

FeaturesSequence of words in the document determine the features in the predictive model.Further processing, such as exponential smoothing of various lengths

23

Documentw1

w2

w3

…wn

Features from Attr DictionaryUD[w1] V1[w1] V2[w1] UD[w2] V1[w2] V2[w2]UD[w3] V1[w3] V2[w3]

…UD[wn] V1[wn] V2[wn]

3K features

Page 24: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Predictive Models

Page 25: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

ComponentsMultiple streaming variable selection

Depth-first, guided selection

Auction frameworkBlend several strategies raw data, calibration, nonlinearity, interactionFormalize external expert knowledge

Statistics: Estimates and standard errorsSandwich estimator for robust SEShrinkage

Sequential testingAlpha investing avoids need for tuning dataMartingale control of expected false discoveries

Or your favorite method (e.g. R package glmnet)25

Page 26: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Based on RegressionFamiliar, interpretable, good diagnosticsRegression has worked well

Predicting rare events, such as bankruptcyCompetitive with random forestFunction estimation, using wavelets and variations on thresholdingTrick is getting the right explanatory variables

Extend to rich environmentsSpatial-temporal dataRetail credit default!! ! ! ! ! MRF, MCMCLinguistics, text miningWord disambiguation, cloze!! ! TF-IDF

Avoid overfitting...26TF-IDF: term frequency-inverse document frequency

frequency in document relative to frequency in corpus MRF: Markov random fields

Page 27: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Lessons from Prior Work“Breadth-first” search

Slow, large memory spaceFixed set of features in searchSevere penalty on largest z-score, sqrt(2 log p)

If most searched features are interactions, then most selected features are interactionsμ»0 and β1, β2 ≠ 0, then X1*X2 ⇒ c + β1X1 + β2X2

Outliers cause problems even with large n

27

Real p-value ≈ 1/1000, but

usual t-statistic ≈ 10

Page 28: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Feature Auction

28

Auction

Expert1

Expert2 ...

ExpertN

α1 α2

αN

Collection of experts bid for the

opportunity to recommend feature

Auction collects winning bid α2

Expert supplies recommended feature Xw

Xw

pw

Expert receives payoff ω if pw ≤ α2

Experts learn if the bid was accepted, not the effect size or pw.

Ymodel

Stat model returns p-value

Page 29: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Experts

29

Auction

X1,X2,X3,...

Elin

Xj

Equad

XjXk

Z1,Z2,Z3,...

Ecal

Scaven

gers

f(ŷ)Eg

g(Xa1)

Xaccepted Xrejected

SubspacesESVD

ERKHSSj

Bj

Source Experts

Elin

Zj

Elag

Zt-s

Enbd

ZN(c)

Page 30: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

ExpertsStrategy for creating sequence of possible explanatory variables.

Embody domain knowledge, science of application.Source experts

A collection of measurements (CCA features)Subspace basis (PCA, RKHS)Multiple smooths of context variablesInteractions between within/between groups

ScavengersInteractions - among features accepted/rejected by modelTransformations- segmenting, as in scatterplot smoothing- polynomial transformations

Calibration30

Page 31: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

CalibrationSimple way to capture global nonlinearity

aka, nonparametric single-index model

Predictor is calibrated if! ! ! ! ! ! E(Ŷ) = YSimple way to calibrate a model is to regression Y on Ŷ2 and Ŷ3 until linear.

31

0.0 0.2 0.4 0.6 0.8 1.0

0.0

0.2

0.4

0.6

0.8

1.0

Calibration

Predicted Values

Y

third

-0.5 0.0 0.5

-1.0

-0.5

0.0

0.5

1.0

Calibration

Predicted Values

Y

First

0.0 0.2 0.4 0.6 0.8 1.0

0.0

0.2

0.4

0.6

0.8

1.0

Calibration

Predicted Values

Y

second

-1.0 -0.5 0.0 0.5 1.0

-1.0

-0.5

0.0

0.5

1.0

Scatterplot

X

Y

Page 32: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Expert WealthExpert gains wealth if feature accepted

Experts have alpha-wealthIf recommended feature is accepted in the model, expert earns ω additional wealthIf recommended feature is refused, expert loses bid

As auction proceeds...Reward experts that offer useful features. These then can afford later bids and recommend more X’sEliminate experts whose features are not useful.

Taxes fund parasites and scavengersContinue control overall FDR

Criticalcontrol multiplicity in a sequence of hypothesesp-values determine useful features

32

Page 33: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Robust Standard Errorsp-values depend on many things

p-value = f(effect size, std error, prob dist)Error structure likely heteroscedasticObservations frequently dependent

DependenceComplex spatial dependence in default ratesDocuments from various news feedsTransfer learningWhen train on observations from selected regions or document sources, what can you infer to others?

What are the right degrees of freedom?Tukey story

33

Page 34: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Sandwich EstimatorUsual OLS estimate of variance

Assume your model is true

Sandwich estimatorsRobust to deviations from assumptions

34

var(b) = (X’X)-1X’E(ee’)X(X’X)-1 = σ2(X’X)-1(X’X) (X’X)-1 = σ2(X’X)-1

heteroscedasticityvar(b) = (X’X)-1X’E(ee’)X(X’X)-1 = (X’X)-1 X’D2X (X’X)-1

diagonal

dependencevar(b) = (X’X)-1X’E(ee’)X(X’X)-1 = σ2(X’X)-1 X’BX (X’X)-1

block diagonal

Essentially the “Tukey method”

Page 35: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Flashback...Heteroscedastic errors

Estimate standard error with outlierSandwich estimator allowing heteroscedastic error variances givesa t-stat ≈ 1, not 10.

Dependent errorsEven more critical to obtain an accurate SENetflix exampleBonferroni (hard thresholding) overfits due to dependence in responses.Credit default modelingEverything seems significant unless incorporate dependence into the calculation of the SE

35

Page 36: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

EstimatorsShrinkage

Two estimates of βj: 0 and bj

Std error determines the amount of shrinkage- Larger the t-statistic, the smaller the shinkageResembles Bayes estimator with Cauchy prior“Smooth” version of hard thresholding

36

-3 -2 -1 0 1 2 3

-3-2

-10

12

3

t-stat,shrunkenestimate

t-stat, LS estimate

Page 37: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Alpha InvestingContext

Test possibly infinite sequence of m hypotheses! ! H1, H2, H3, … Hm … obtaining p-values p1, p2, ...Order of tests can depend prior outcomes

ProcedureStart with an initial alpha wealth W0 = αInvest wealth 0 ≤ αj ≤ Wj in the test of Hj

Change in wealth depends on test outcomeω ≤ α denotes the payout earned by rejecting

37

Wj - Wj-1 =ω if pj ≤ αj

-αj if pj > αj{

Page 38: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Martingale ControlProvides uniform control of the expected false discovery rate. At any stopping time during testing, martingale argument shows

Flexibility in choice of how to invest alpha-wealth in test of each hypothesis

Invest more when just reject if suspectthat significant results cluster.Universal investing strategies

Avoids computing all p-values in advance

38

E(#false rejects)E(#rejects)+1

supθ

≤ α

Page 39: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Multiple TestingOther methods are special cases

Note: alpha-investing does not require the full set of p-values or estimates at the start.

Bonferroni test of H1,…,HmSet initial W0 = α and reward to ω = 0.05.Bid αj = α/m

Step-down test of Benjamini and HochbergSet initial W0 = α and reward to ω = 0.05.Test H1,...Hm at fixed level α/mIf none reject -> finished.If one rejects, earn α = 0.05 for next roundTest next round conditionally on pj > α/m-> continue with remaining hypotheses.

39

Page 40: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Example…Back to text processing

Page 41: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Named Entity Results Model

Approximate max entropy classifierFancy name for multinomial logit

Other predictive models can be used

DataPortion of the ConLL03 dataTraining and test subsets

Dictionaryd=50,000 wordsExponential smooths of content featuresInteractions

Precision and recall about 0.85

41

Page 42: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Auction Run

42

First 2,000 rounds of auction modeling.

0 500 1000 1500

Calibration rounds not shownAuction Round

P-Value

-9

.0000001

.00001

.001

.05

.5

Alpha-Wealth

11024.2

12703.5

14382.9

CVSS

Page 43: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

What are the predictors?Interactions

Combinations of canonical variables

Principal components of factorsCombinations of skipped featuresRKHS finds some nonlinear combinations

Calibration adjustmentsSimple method to estimate single-index model! ! ŷ = g(b0 + b1 X1 + … + bkXk)Estimate g cheaply by building a nonlinear regression of y on linear ŷ.

43

Page 44: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Closing Comments

Page 45: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Next StepsText

Incorporate features from other methodsUnderstanding the CCA Other “neighborhood” features

TheoryDevelop martingale that controls expected loss.Adapt theory from the “nearly black” world of modern statistics to “nearly white” world of text

ComputingMulti-threading is necessary to exploit trend toward vast number of cores in CPUMore specialized matrix code

45

Page 46: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

Linguistics ≈ Spatial TS

Predict word in new documents, different authorsLatent structure associated with corpus

Neighborhoods: nearby words, sentences

Vast possible corpus

Sparse

Predict rates in same locations, but changing economic conditionsLatent temporal changes as economy evolves

Neighborhoods: nearby locations, time periods

70 quarters, 3000 counties. Possible to drill lower.

May be sparse

46

Text Credit default

Page 47: Text Mining Using Linear Models of Latent Statesstine/research/siam_dm_2011.pdf · Predictive Models. Wharton Department of Statistics Components Multiple streaming variable selection

Wharton Department of Statistics

ReferencesFeature auction

www-stat.wharton.upenn.edu/~stine

Alpha investing“α-investing: a procedure for sequential control of expected false discoveries”, JRSSB. 2008

Streaming variable selection“VIF regression”, JASA. 2011

Linear structure of HMM“A spectral algorithm for learning hidden Markov models”, Hsu, Kakade, Zhang, TTI. 2008

Random projections“Finding structure with randomness”, Halko, Martinsson, and Tropp. 2010

47Thanks!


Recommended