+ All Categories
Home > Documents > Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data...

Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data...

Date post: 14-Mar-2020
Category:
Upload: others
View: 7 times
Download: 0 times
Share this document with a friend
36
Machine Learning Srihari Kernel Methods Sargur Srihari [email protected] 1
Transcript
Page 1: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernel Methods

Sargur [email protected]

1

Page 2: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Topics in Kernel Methods

1. Linear Models vs Memory-based models2. Stored Sample Methods3. Kernel Functions

• Dual Representations• Constructing Kernels

4. Extension to Symbolic Inputs5. Fisher Kernel

2

Page 3: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Linear Models vs Memory-based models

• Linear parametric models for regression and classification have the form y(x,w)• During learning we either get a maximum likelihood

estimate of w or a posterior distribution of w• Training data is then discarded• Prediction based only on vector w• This is true of Neural networks as well

• Memory-based methods keep the training samples (or a subset) and use them during the prediction phase

3

Page 4: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Memory-Based Methods

• Training data points are used in prediction• Examples of such methods

• Parzen probability density model• Linear combination of kernel functions centered on each training data point

• Nearest neighbor classification• These are memory-based methods• They require a metric to be defined

• Fast to train, slow to predict

4

Page 5: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariKernel Functions• Linear models can be re-cast

• into equivalent dual where predictions are based on kernel functions evaluated at training points

• Kernel function is given byk (x,x�) = ϕ(x)T ϕ(x�)• where ϕ(x) is a fixed nonlinear mapping (basis function)

• Kernel is a symmetric function of its argumentsk (x,x�) = k (x�,x)

• Kernel can be interpreted as similarity of x and x�• Simplest is identity mapping in feature space ϕ(x) = x• In which case k (x,x�) = xTx�

• Called Linear Kernel 5

Page 6: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariKernel Trick• Formulated as inner product allows extending

well-known algorithms • by using the kernel trick

• Basic idea of kernel trick • If an input vector x appears only in the form of

scalar products then we can replace scalar products with some other choice of kernel

• Used widely• in support vector machines• in developing non-linear variant of PCA• In kernel Fisher discriminant 6

Page 7: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Other Forms of Kernel Functions

• Function of difference between argumentsk (x,x�) = k (x-x�)

• Called stationary kernel since invariant to translation

• Radial basis functionsk (x,x�) = k (||x-x�||)• Depends on magnitude of distance between arguments

• Note that the kernel function is scalar while x is M-dimensional

7

For these to be valid kernel functions they should be shown to have the propertyk (x,x�) = φ (x)T φ(x�)

Page 8: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Dual Representation

• Linear models for regression/classification can be reformulated in terms of dual representation• In which kernel function arises naturally• Plays important role in SVMs

• Consider linear regression model• Parameters from minimizing sum-of-squares

• Minimum obtained by setting gradient of J(w) wrt w equal to zero

J(w) =12

wTφ(xn ) − tn{ }2

+λ2n=1

N

∑ wTw

where w = (w0,.., wM-1)T , φ = (φ0 ,..φM-1)T

we have N samples {x1,..xN}λ is the regularization coefficient

ϕ is the set of Mbasis functionsor feature vector

Page 9: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Solution for w: linear combination of φ (xn)• Equating derivative J(w) wrt w to zero we get

• Solution for w: linear combination of φ (xn) whose coefficients are functions of w where• Φ is design matrix whose nth row is given by φ (xn)T

• Vector a=(a1,..,aN)T has the definition

w = −1λ

wTφ(xn ) − tn{ }n=1

N

∑ φ(xn )

= ann=1

N

∑ φ(xn )

=ΦTa

a

n=−

wTφ(xn)− t

n{ }

Φ =

φ0(x1) . . φM −1(x1). .

φ0(xn ).

. . φM −1(xn ).

φ0(xN ) . . φM −1(xN )

%

&

' ' ' ' '

(

)

* * * * *

is a N ×M matrix

Page 10: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Transformation from w to a

• Thus we have • Instead of working with parameter vector w we

can reformulate least squares algorithm in terms of parameter vector a• giving rise to dual representation

• Although the definition of a still includes wIt can be eliminated by the use of the kernel function

10

w = ΦTa

a

n=−

wTφ(xn)− t

n{ }

Page 11: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Gram Matrix and Kernel Function

• Gram matrix K=ΦΦT is N x N matrix• with elements

Knm= ϕ (xn)Tϕ (xm)=k(xn,xm)• where kernel function k (x,x�) = ϕ (x)T ϕ (x�)

• Notes: • Φ is NxM and K is NxN• K is a matrix of similarities of pairs of samples (thus it is symmetric) 11

Gram Matrix Definition:Given N vectors, it is thematrix of all inner products

K =

k(x1,x

1) . . k(x

1,x

N)

.

.k(x

N,x

1) k(x

N,x

N1)

⎢⎢⎢⎢⎢⎢⎢

⎥⎥⎥⎥⎥⎥⎥

Note: N �M times M � N

Page 12: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariError Function in Terms of Gram Matrix• Error Function is

• Substituting w = ΦTa into J(w) gives

where t = (t1,..,tN)T

• Error function is written in terms of Gram matrix as

• Solving for a by combining w=ΦTa and

a =(K +λIN)-1t

Solution for a can be expressed as a linear combination of elements of ϕ(x) whose coefficients are entirely in terms of kernel k(x,x�) from which we can recover original formulation in terms of parameters w

J(a) =

12aTKKa−aTKt +

12

tTt +λ2aTKa

J(w) =

12

wTφ(xn)− t

n{ }2

+λ2n=1

N

∑ wTw

J(w) =

12aTΦΦTΦΦTa−aTΦΦTt +

12

tTt +λ2aTΦΦTa

an = −1λ

wTφ(xn ) − tn{ }

Page 13: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariPrediction Function

• Prediction for new input x• We can write a =(K +λIN)-1t by combining w =ΦTa and

• Substituting back into linear regression model,

• Prediction is a linear combination of the target values from the training set.

y(x) = wTφ(x)

= aTΦφ(x)

= k(x)T(K +λIN)−1t where k(x) has elements k

n(x) = k(x

n,x)

an = −1λ

wTφ(xn ) − tn{ }

Page 14: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Advantage of Dual Representation• Solution for a is entirely in terms of k(x,x�)

• From a we can recover w using w = Φta• In parametric formulation, solution is

• We invert M xM matrix, since Φ is N x M

• In dual formulation, solution is a =(K +λIN)-1t

• We are inverting an N x N matrix• An apparent disadvantage, since N >>M

• Advantage of dual: work with kernel k(x,x�), so:• avoid working with a feature vector ϕ(x) and • problems associated with very high or infinite

dimensionality of x 14

wML= (ΦTΦ)−1ΦTt

Page 15: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Constructing Kernels

• To exploit kernel substitution need valid kernels• Two methods:

• (1) from ϕ (x) to k(x,x’)• (2) from k(x,x’) to ϕ (x)

• First Method• Choose ϕ (x) and use it to find corresponding kernel

• where ϕi(x) are basis functions• One-dimensional input space is shown next

15

k(x,x ') = φ(x)Tφ(x ')

= φi(x)φ

i(x ')

i=1

M

Page 16: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernel Functions from basis functions

16

Polynomials Gaussian Logistic Sigmoid

Red cross is x�

BasisFunctionsϕi(x)

KernelFunctionsk(x,x�) = ϕ (x)Tϕ (x’)

One-dimensional input space

Page 17: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Method 2: Direct Construction of Kernels• Kernel has to correspond to a scalar product in

some (perhaps infinite dimensional) space• Consider kernel function k(x,z) = (xTz)2

• In 2-D space

• Feature mapping takes the form• Comprises of all second order terms with a specific weighting• Inner product needs computing 6 feature values, 9 multiplications

• Kernel function k(x,z) has 2 multiplies and a squaring• For(xTz+c)2 we get constant, linear, second order terms• For (xTz+c)M we get all powers of x (monomials)

17

k(x,z) = (xTz)2 = (x1z

1+ x

2z

2)2

= x12z

12 + 2x

1z

1x

2z

2+ x

22z

22

= (x12, 2x

1x

2,x

22)(z

12, 2z

1z

2,z

22)T

= φ(x)Tφ(z)

φ(x) = (x12, 2x

1x

2,x

22)

Page 18: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Testing whether a function is a valid kernel

• Without constructing ϕ(x), Necessary and sufficient condition for k(x,x�) to be a kernel is • Gram matrix K, with elements k(xn,xm) is positive

{xn}semi-definite for all possible choices of the set• Positive semi-definite is not the same thing as a matrix whose elements are non-

negative• It means

• Mercer�s theorem: any continuous, symmetric, positive semi-definite kernel function k(x, y) can be expressed as a dot product in a high-dimensional space

• New kernels can be constructed from simpler kernels as building blocks 18

zTKz ≥ 0 for non - zero vectors z with real entries

i.e., Knmm∑

n∑ znzm ≥ 0 for any real numbers zn , zm

Page 19: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Techniques for Constructing Kernels• Given valid kernels k1(x,x�) and k2(x,x�) the following new

kernels will be valid1. k(x,x�) =ck1(x,x�)2. k(x,x�)=f(x)k1(x,x�)f(x�)3. k(x,x�)=q(k1(x,x�))4. k(x,x�)=exp(k1(x,x�))5. k(x,x�)=k1(x,x�)+k2(x,x�)6. k(x,x�)=k1(x,x�)k2(x,x�)7. k(x,x�)=k3(f(x).f(x�))8. k(x,x�)=xTAx�9. k(x,x�)=ka(xa,xb�)+kb(xb,xb�)10. k(x,x�)=ka(xa,xa�)kb(xb,xb�)

Wheref (.) is any function

q(.) is a polynomial with non-negative coefficients

f(x) is a function from x to RM

k3 is a valid kernel in RM

A is a symmetric positive semidefinite matrix

xa and xb are variables with x=(xa, xb)ka and kb are valid kernel functions

Page 20: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernels for specific applications

• Requirements for k(x,x�)• It is symmetric• Its Gram matrix is positive semidefinite• It expresses the appropriate similarity between x

and x� for the intended application

20

Page 21: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariGaussian Kernel• Commonly used kernel is

k(x,x�) = exp (-||x-x�||2/2σ2)

• It is seen as a valid kernel by expanding square||x-x�||2 = xTx + (x�)Tx� -2xTx�• To give

k(x,x�) = exp (-xTx/2σ2) exp (-xTx�/σ2) exp (-(x�)Tx�/2σ2)

• From kernel construction rules 2 and 4• together with validity of linear kernel k(x,x�)=xTx�

• Can be extended to non-Euclidean distancesk(x,x�) = exp {(-1/2σ2)[κ(x,x�)+κ (x�,x�)-2κ (x,x�)]}

21

Page 22: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

22

Use of a kernel method

2-D space x

Infinite dimensional space ϕ(x)Induced by a Gaussian kernel

3-D space ϕ(x)

Oil slick 2D- imagethat looks like a feature space x

Page 23: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Extension of Kernels to Symbolic Inputs• Important contribution of kernel viewpoint:

• Inputs are symbolic, not vectors of real numbers• Kernels defined for graphs, sets, strings, text documents

• If A1 and A2 are two subsets of objects• A simple kernel is

• where |A | indicates no of elements in A• A valid kernel since it can be shown to correspond to an

inner product in a feature space

k(A1,A

2) = 2|A1∩A2|

A={1,2,3,4,5}

A1={2,3,4,5}, A2={1,2,4,5}, A1∩A2={2,4,5}

Hence k(A1,A2)=8

Example feature vectors:

ϕ(A1)=[2 2 2 2] and ϕ(A2) =[1 1 1 1]

such that

ϕ(A1)Tϕ(A2)=8

Page 24: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernels for Complex Objects• ML methods studied so far require input

represented as a fixed-size feature vector xiε RD• For certain objects it is not clear how best to

represent them as feature vectors, E.g., • Text or protein sequence of variable length• Molecular structure with complex 3-D geometry• Evolutionary tree which has variable size and shape

• Solutions• Define generative model, whose parameters are

features. Plug these features into standard models• Measure similarity between objects which does not

require preprocessing into feature vector format 24

Page 25: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernels for Comparing Documents• In IR and document classification, need a way

of comparing documents xi and xi’• Bag of words representation: xij is the no of times

word j occurs in document i• Cosine similarity

• It measures cosine of angle between xi and xi’• Since xi is a count vector (and hence non-negative) the

cosine similarity is between 0 and 1• Where 0 means the vectors are orthogonal and therefore have no

words in common

25

k(x

i,x

i ') =

xiTx

i '

||xi||

2||x

i '||

2

xi=[xi1,..,xiD]

Page 26: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Disadvantage of Cosine Similarity

• It does not work well for two reasons1. Stop words• If xi has any word in common with xi’ it is deemed similar,

even though some words such as the, and occur commonly in many documents and are not discriminative

2. High frequency words in a document• If a discriminatory word occurs many times in a document,

the similarity is artificially boosted• Even though word usage would be bursty, i.e., once a word is used

in a document it is very likely to be used again

26

Page 27: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

TF-IDF• Cosine similarity performance can be improved

with some preprocessing• Use feature vector called TF-IDF representation

• Term frequency-Inverse document frequency• TF is a log-transform of the count• IDF is defined as

• Where N is the total no of documents• Denominator counts how many documents contain term j

• Finally we define

• We use this inside cosine similarity measure

• Where ϕ (x)=tf-idf(x) 27

tf(x

ij)! log(1+ x

ij)

idf(j)! logN

1+ I(xij

> 0)i=1

N∑

tf-idf(xi)! [tf(x

ij)× idf(j)]

j=1V

k(x

i,x

i ') =

φ(xi)Tφ(x

i ')

||φ(xi) ||

2||φ(x

i ') ||

2

xij is the no of times word j occurs in document i

Page 28: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

String Kernels

• Real power of kernels: inputs are structured • Compare two variable-length strings

• Strings x, and x’ of length D, D’ defined over A• E.g., two amino acid sequences defined over 20-letter

alphabet A={A,R,N,D,C,E,Q,G,H,I,L,K,M,F,P,S,T,W,Y,V}• x is the sequence of length 110IPTSALVKETLALLSTHRTLLIANETLRIPVPVHKN……VNQFLDYLQEFLGVMNTEWI

• x’ is a string of length 153• PHRRDLCSRSIWLARKIRSDLTALTESYVKHQGLWSELTEAERLQENLQAYRTFHVLLA……..

• The strings have the substring LQE in common

• We can define the similarity of two strings to be the no of substrings they have in common

• Definition given next28

Page 29: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Mercer Kernel• If s is a substring of x we can write x=usv for

some (possibly empty) strings u,s and v• ϕs(x) is no of times substring s appears in x• Define kernel between two strings x and x’ as

• Where ws≥0 and A* is the set of all strings from alphabet A(known as Kleene * operator)

• Can be computed in O(|x|+|x’|) time

• Cases of interest: • ws=0 for |s|>1 we get bag-of-characters kernel– defines ϕ(x) as no of times each character in A occurs in x

• If we require s to be bordered by white space we get bag-of- words kernel 29

k(x,x ') = w

ss∈A*∑ φ

s(x)φ

s(x ')

Page 30: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Combining Discriminative and Generative Models

• Generative models deal naturally with missing data and with HMM of varying length

• Discriminative models such as SVM have better performance

• Can use a generative model to define a kernel and use kernel in discriminative approach

30

Page 31: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernels based on Generative Models

• For a model p(x) we define a kernel byk (x,x�) = p(x) p(x�)• A valid kernel since an inner product in 1-dimensional

feature space defined by the mapping p(x)

• Two inputs x and x� are similar if they have high probabilities

31

Page 32: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Kernel Functions based on Mixture Densities• Extension to sums/products of distributions

• where p(i) are positive weighting coefficients• A valid kernel based on two rules of kernel construction: k(x,x�) =ck1(x,x�) and k(x,x�)=k1(x,x�)+k2(x,x�)

• Two inputs x and x� will give a large value of k , and hence appear similar, if they have a significant probability under a range of different components

• Taking the limit to infinite sum

• where z is a continuous latent variable32

k(x,x ') = p(x | i)p(x ' | i)p(i)

i∑

k(x,x ') = p(x | z)p(x ' | z)p(z)∫ dz

Page 33: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariKernels for Sequences

• Data consists of ordered sequences of length LX={x1,..,xL}

• Generative model for sequences is HMM• Hidden states Z={z1,..,zL}

• Kernel Function for measuring similarity of sequences X and X� is

• Both observed sequences are generated by same hidden sequence Z

33

k(X,X ') = p(X | Z)p(X ' | Z ')p(Z)

Z∑

Page 34: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariFisher Kernel• Alternative technique for generative models

• In document retrieval, protein sequences• Consider parametric generative model p(x|θ)

where θ denotes vector of parameters• Goal: find kernel that measures similarity of two

vectors x and x� induced by the generative model• Define Fisher score as gradient wrt θ

• Fisher Kernel is

where F is the Fisher information matrix 34

g(θ,x) =∇θ ln p(x |θ)

k(x,x ') = g(θ, x)T F−1g(θ, x’)

F = Ex g(θ,x)g(θ,x)T[ ]

A vector of same dimensionality as θ

Fisher score is moregenerally the gradientof the log-likelihood

Page 35: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning Srihari

Fisher Information Matrix• Presence of Fisher information matrix causes

kernel to be invariant under non-linear parametrization of the density model θàψ(θ)• In practice, infeasible to evaluate Fisher Information

Matrix. Instead use the approximation

• This is the covariance matrix of the Fisher scores• So the Fisher kernel corresponds to whitening of the

Fisher scores• More simply omit F and use non-invariant kernel

35

F ≈1N

g(θ,xn )g(θ,xn )T

n=1

N

k(x,x') = g(θ,x)T F−1g(θ,x’)

k(x,x') = g(θ,x)T g(θ,x')

Page 36: Kernel Methods - University at BuffaloMachine Learning Srihari Memory-Based Methods •Training data points are used in prediction •Examples of such methods •Parzen probability

Machine Learning SrihariSigmoidal Kernel• A link between SVMs and neural network

k (x,x�) = tanh (axTx� + b)• Its Gram matrix is not positive semidefinite• But used in practice because it gives SVMs a

superficial resemblance to neural networks• Bayesian neural network with an appropriate

prior reduces to a Gaussian process• Provides a deeper link between neural networks

and kernel methods

36


Recommended