+ All Categories
Home > Documents > Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop...

Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop...

Date post: 27-Mar-2018
Category:
Upload: duongkiet
View: 221 times
Download: 0 times
Share this document with a friend
50
Workshop report 1. Daniels report is on website 2. Dont expect to write it based on listening to one project (we had 6 only 2 was sufficient quality) 3. I suggest writing it on one presentation. 4. Include figures (from a related paper or their presentation) 5. Include references
Transcript
Page 1: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Workshopreport1. Danielsreportisonwebsite2. Don’texpecttowriteitbasedonlisteningtooneproject(wehad6

only2wassufficientquality)3. Isuggestwritingitononepresentation.4. Includefigures(fromarelatedpaperortheirpresentation)5. Includereferences

Page 2: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

May8,CODYMachineLearningforfindingoil,focusingon1)robustseismicdenoising/interpolationusingstructuredmatrixapproximation2)seismicimageclusteringandclassification,usingt-SNE(t-distributedstochasticneighborembedding)andCNN.Weichang Li,Goup LeaderAramco,Houston.

May10,ClassHWFirstdistributionoffinalprojects.Oceanacousticsourcetracking.Finalprojects.Finalprojectisthemaingoalinlastmonth. BishopCh 9Mixturemodels

May15,CODY SeismologyandMachineLearning,DanielTrugman (halfclass),ch 8GraphicalmodelsMay17,ClassHWch 8May22,Dictionarylearning,MikeBianco(halfclass),BishopCh 13May24,ClassHWBishopCh 13MAY30CODYMay31,NoClass.Workshop, BigDataandTheEarthSciences:GrandChallengesWorkshopJune5,Discussworkshop,ch13.Spiess Hallopenforprojectdiscussion11am-.June7,Workshopreport.NoclassJune12Spiess Hallopenforprojectdiscussion9-11:30amand2-7pmJune16Finalreportdelivered.Beertime

ForfinalprojectdiscussioneveryafternoonMarkandIwillbeavailable

Chapter13Sequentialdata

Page 3: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ProblemsOceansourcetrackingXRe-implementSourceLocalizationinanOceanWaveguideusingSupervisedMachineLearningX-rayspectrumabsorptioninterpretationusingNNNeuraldecodingPlanktonTransferlearninganddeepfeatureextractionforplanktonicimagedatasetsSpeakertaggerCoralResturantAmazonrainforest(Kaggle)Myshake SeismicHigh-precisionindoorpositioningframeworkformostwifi-enableddevices

PleaseaskquestionsMarkandIavailableallafternoons.Justcomeoremailfortimeslots.Spiess hall330isopenMonday5and12June.IfinterestedIcanbookitatothertimes

ReportRatherconcisethanlong.Largergroupcandomore.Startwithsomeverysimpleexample.Toshowyourideaandthatitisworking.EndwithshowingtheadvancedabilitiesSeveralfigures.Equationsarenice.

DeliveryZipfile(Friday16)Maincode(notall).Itshouldbeabletorun.Report(pdfpreferred).

FinalReport

Page 4: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

PATTERNRECOGNITIONAND MACHINELEARNINGCHAPTER8:GRAPHICALMODELS

Page 5: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

10.1. Introduction 309

4 5

2 3

1

(a)

4 5

2 3

1

(b)

Figure 10.1 (a) A simple DAG on 5 nodes, numbered in topological order. Node 1 is the root, nodes 4 and5 are the leaves. (b) A simple undirected graph, with the following maximal cliques: {1, 2, 3}, {2, 3, 4},{3, 5}.

10.1.4 Graph terminology

Before we continue, we must define a few basic terms, most of which are very intuitive.A graph G = (V , E) consists of a set of nodes or vertices, V = {1, . . . , V }, and a set

of edges, E = {(s, t) : s, t ∈ V}. We can represent the graph by its adjacency matrix, inwhich we write G(s, t) = 1 to denote (s, t) ∈ E , that is, if s → t is an edge in the graph.If G(s, t) = 1 iff G(t, s) = 1, we say the graph is undirected, otherwise it is directed. Weusually assume G(s, s) = 0, which means there are no self loops.

Here are some other terms we will commonly use:

• Parent For a directed graph, the parents of a node is the set of all nodes that feed into it:pa(s) ! {t : G(t, s) = 1}.

• Child For a directed graph, the children of a node is the set of all nodes that feed out of it:ch(s) ! {t : G(s, t) = 1}.

• Family For a directed graph, the family of a node is the node and its parents, fam(s) ={s} ∪ pa(s).

• Root For a directed graph, a root is a node with no parents.• Leaf For a directed graph, a leaf is a node with no children.• Ancestors For a directed graph, the ancestors are the parents, grand-parents, etc of a node.

That is, the ancestors of t is the set of nodes that connect to t via a trail: anc(t) ! {s : s ❀t}.

• Descendants For a directed graph, the descendants are the children, grand-children, etc ofa node. That is, the descendants of s is the set of nodes that can be reached via trails froms: desc(s) ! {t : s ❀ t}.

• Neighbors For any graph, we define the neighbors of a node as the set of all immediatelyconnected nodes, nbr(s) ! {t : G(s, t) = 1 ∨ G(t, s) = 1}. For an undirected graph, we

Page 6: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

310 Chapter 10. Directed graphical models (Bayes nets)

write s ∼ t to indicate that s and t are neighbors (so (s, t) ∈ E is an edge in the graph).• Degree The degree of a node is the number of neighbors. For directed graphs, we speak of

the in-degree and out-degree, which count the number of parents and children.• Cycle or loop For any graph, we define a cycle or loop to be a series of nodes such that

we can get back to where we started by following edges, s1 − s2 · · ·− sn − s1, n ≥ 2. If thegraph is directed, we may speak of a directed cycle. For example, in Figure 10.1(a), there areno directed cycles, but 1→ 2→ 4→ 3→ 1 is an undirected cycle.

• DAG A directed acyclic graph or DAG is a directed graph with no directed cycles. SeeFigure 10.1(a) for an example.

• Topological ordering For a DAG, a topological ordering or total ordering is a numberingof the nodes such that parents have lower numbers than their children. For example, inFigure 10.1(a), we can use (1, 2, 3, 4, 5), or (1, 3, 2, 5, 4), etc.

• Path or trail A path or trail s ❀ t is a series of directed edges leading from s to t.• Tree An undirected tree is an undirectecd graph with no cycles. A directed tree is a DAG in

which there are no directed cycles. If we allow a node to have multiple parents, we call it apolytree, otherwise we call it a moral directed tree.

• Forest A forest is a set of trees.• Subgraph A (node-induced) subgraph GA is the graph created by using the nodes in A and

their corresponding edges, GA = (VA, EA).• Clique For an undirected graph, a clique is a set of nodes that are all neighbors of each

other. A maximal clique is a clique which cannot be made any larger without losing theclique property. For example, in Figure 10.1(b), {1, 2} is a clique but it is not maximal, sincewe can add 3 and still maintain the clique property. In fact, the maximal cliques are asfollows: {1, 2, 3}, {2, 3, 4}, {3, 5}.

10.1.5 Directed graphical models

A directed graphical model or DGM is a GM whose graph is a DAG. These are more commonlyknown as Bayesian networks. However, there is nothing inherently “Bayesian” about Bayesiannetworks: they are just a way of defining probability distributions. These models are also calledbelief networks. The term “belief” here refers to subjective probability. Once again, there isnothing inherently subjective about the kinds of probability distributions represented by DGMs.Finally, these models are sometimes called causal networks, because the directed arrows aresometimes interpreted as representing causal relations. However, there is nothing inherentlycausal about DGMs. (See Section 26.6.1 for a discussion of causal DGMs.) For these reasons, weuse the more neutral (but less glamorous) term DGM.

The key property of DAGs is that the nodes can be ordered such that parents come beforechildren. This is called a topological ordering, and it can be constructed from any DAG. Givensuch an order, we define the ordered Markov property to be the assumption that a node onlydepends on its immediate parents, not on all predecessors in the ordering, i.e.,

xs ⊥ xpred(s)\pa(s)|xpa(s) (10.4)

where pa(s) are the parents of node s, and pred(s) are the predecessors of node s in theordering. This is a natural generalization of the first-order Markov property to from chains togeneral DAGs.

Page 7: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Three types of graphical model

Directed graphs – useful for designing models

Undirected graphs – good for some domains, e.g. computer vision

Factor graphs – useful for inference and learning

Page 8: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

BayesianNetworks(BayesNets)orDirectedgraphicalmodel(DGM)Decomposition

Page 9: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

DirectedGraphsorBayesianNetworks

GeneralFactorization

Page 10: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

BayesianCurveFitting(1)

Polynomial

Plate

Page 11: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

BayesianCurveFitting(3)Inputvariablesandexplicithyperparameters

Conditionondata

Page 12: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

BayesianCurveFitting —Prediction

Predictivedistribution:

where

Page 13: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

GenerativeModelsCausalprocessforgeneratingimages

Page 14: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

DiscreteVariables(1)Generaljointdistribution:K2-1 parameters

Independentjointdistribution:2(K-1) parameters

GeneraljointdistributionoverM variables:KM -1parametersM-nodeMarkovchain:K-1 + (M-1) K(K-1) parameters

Page 15: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

DiscreteVariables:BayesianParameters(1)

Page 16: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

DiscreteVariables:BayesianParameters(2)

Sharedprior

Page 17: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ParameterizedConditionalDistributions

If arediscrete,K-statevariables,

ingeneralhasO(KM) parameters.

Theparameterizedform

requiresonlyM + 1 parameters

Page 18: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ConditionalIndependencea isindependentofb givenc

Equivalently

Notation

Page 19: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ConditionalIndependence:Example1

Page 20: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ConditionalIndependence:Example2

Page 21: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Note:thisistheoppositeofExample1,withc unobserved.

ConditionalIndependence:Example3

Page 22: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

D-separation• A,B,andCarenon-intersectingsubsetsofnodesinadirectedgraph.• pathfromAtoBisblockedifitcontainsanodesuchthateither

a) thearrowsonthepathmeeteitherhead-to-tail ortail-to-tail atthenode,andthenodeisinthesetC,or

b) thearrowsmeethead-to-head atthenode,andneitherthenode,noranyofitsdescendants,areinthesetC.

• IfallpathsfromAtoBareblocked,Aissaidtobed-separatedfromBbyC.Thenthejointdistributionoverallvariablessatisfies.

D-separation:Example

Page 23: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

MarkovRandomFieldsorUndirectedGraphs

Page 24: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

CliquesandMaximalCliques

Clique

MaximalClique

Page 25: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

JointDistributionWhereisthepotentialovercliqueC and

isthenormalizationcoefficient;note:M K-statevariables® KM termsinZ.

EnergiesandtheBoltzmanndistribution

Page 26: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Illustration:ImageDe-Noising

NoisyImage RestoredImage(ICM) RestoredImage(Graphcuts)

Page 27: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ConvertingDirectedtoUndirectedGraphs(1)

Page 28: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

ConvertingDirectedtoUndirectedGraphs(2)

Additionallinks

Page 29: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Directedvs.UndirectedGraphs(2)

Page 30: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

InferenceinGraphicalModels

Page 31: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

InferenceonaChain

Page 32: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

InferenceonaChain

Page 33: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

InferenceonaChain

Page 34: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

InferenceonaChain

Tocomputelocalmarginals:• Computeandstoreallforwardmessages,.• Computeandstoreallbackwardmessages,.• ComputeZ atanynodexm

• Compute

forallvariablesrequired.

Page 35: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

TreesUndirectedTree DirectedTree Polytree

Page 36: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Factorization

Directed graphs: Undirected graphs: Both have the form of products of factors:

Page 37: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

FactorGraphs

More verbose!

Page 38: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

fromDirectedGraphstoFactorGraphs

Page 39: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

FactorGraphsfromUndirected Graphs

Page 40: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

INFERENCE

Page 41: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

TheSum-ProductAlgorithm(1)Objective:

i. toobtainanefficient,exactinferencealgorithmforfindingmarginals;ii. insituationswhereseveralmarginalsarerequired,toallowcomputationsto

besharedefficiently.

Keyidea:DistributiveLaw

Efficient inference

7 versus 3 operations

Page 42: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

The Sum-Product Algorithm

𝑥 𝑢 𝑤 𝑦

𝑧 𝑓1(𝑢, 𝑤) 𝑓2(𝑤, 𝑥)

𝑓4(𝑥, 𝑧)

𝑓3(𝑥, 𝑦)

Page 43: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

What if the messages are intractable?

True distribution Monte Carlo Variational Message Passing

Expectation propagation

Page 44: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Learning is just inference!

Page 45: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

TheSum-ProductAlgorithm(2)

Page 46: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

TheSum-ProductAlgorithm(3)

Page 47: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

TheSum-ProductAlgorithm(4)

Page 48: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

TheSum-ProductAlgorithm(5)Initialization

Tocomputelocalmarginals:• Pickanarbitrarynodeasroot• Computeandpropagatemessagesfromtheleafnodestotheroot,storing

receivedmessagesateverynode.• Computeandpropagatemessagesfromtheroottotheleafnodes,storing

receivedmessagesateverynode.• Computetheproductofreceivedmessagesateachnodeforwhichthe

marginalisrequired,andnormalizeifnecessary.

Page 49: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Sum-Product:Example(1)

Page 50: Workshop report - noiselab.ucsd.edunoiselab.ucsd.edu/ECE285/Lecture13.pdf · June 7, Workshop report. No class ... Bayesian Curve Fitting (3) Input variables and explicit hyperparameters

Sum-Product:Example(2)and(3)


Recommended