+ All Categories
Home > Documents > Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of...

Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of...

Date post: 25-Jan-2020
Category:
Upload: others
View: 0 times
Download: 0 times
Share this document with a friend
23
Inferring the Structure of Probabilistic Graphical Models for Efficient Natural Language Understanding Istvan Chung Oron Propp Mentor: Dr. Thomas Howard 1 1 Robust Robotics Group MIT CSAIL Fourth Annual MIT PRIMES Conference 18 May 2014
Transcript
Page 1: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Inferring the Structure of Probabilistic Graphical Modelsfor Efficient Natural Language Understanding

Istvan Chung Oron ProppMentor: Dr. Thomas Howard1

1Robust Robotics GroupMIT CSAIL

Fourth Annual MIT PRIMES Conference18 May 2014

Page 2: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Introduction

Existing interfaces for controlling robots are specialized and difficultto useIt would be much easier to control robots using natural languagecommandsExisting natural language interfaces do not scale well with thecomplexity of the environment

Page 3: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Probabilistic Graphical Models

ProbabilisticGraphical

Model

world

command

grounding

Example

{WorldObject(0, ‘robot’), WorldObject(1, ‘crate’),

WorldObject(2, ‘box’)} + “approach the box” →Constraint(WorldObject(0), WorldObject(2), ‘near’)

Page 4: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Grammar

It doesn’t make sense to view the input as a monolithic block of text

It is more meaningful to understand the input with its grammaticalstructure

A grammar is used to assign meaning to the words

VP → VB NPVP → VB NP PPVP → VB PPNP → DT NNNP → NP PPPP → IN NPVB → “approach”, “land”, “fly”DT → “a”, “the”NN → “box”, “chair”, “table”IN → “near”, “far”, “to”

Page 5: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Parse Tree

VP

VB

NP

NP

DT NN

PP

IN

NP

DT NN

approach the box near the chair

Figure: Parse tree for “approach the box near the chair”

Page 6: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Parse Ambiguity

Some sentences are ambiguous

VP

VB

NP

DT NN

PP

IN

NP

DT NN

approach the box near the chair

Figure: Alternate parse tree for “approach the box near the chair”

Page 7: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

CYK Chart Parser

The CYK Parsing algorithm [4, 5, 6] accomplishes this task in O(n3)time.

All possible parses of an ambiguous sentence are returned

VP

NP/X0

PP

NP NP

VB DT NN IN DT NN

approach the box near the chair

Page 8: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Generalized Grounding Graph

Comprised of “factors” which relate groundings, correspondences, andphrases, and are represented by log-linear modelsGrounding each phrase depends on the groundings of the child phrases

true true true true true

approach the box near the chair

λ1 λ2 λ3 λ4 λ5

φ1 φ2 φ3 φ4 φ5

γ1

1.36× 10390

γ2 γ3

36γ4

32γ5

36

f1 f2 f3 f4 f5

Figure: Generalized grounding graph for “approach the box near the chair”

[2] S. Tellex, T. Kollar, S. Dickerson, M. Walter, A. Banerjee, S. Teller, and N. Roy,Approaching the Symbol Grounding Problem with Probabilistic Graphical Models. 2013.

Page 9: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Log-linear Model

Log-linear models [1] are used to assign a score to a grounding givensome input.

This is done using a set of features

Features evaluate aspects of the input and grounding

Scoring function

p(c | x , y ; v) =exp(v · f(x , y , c))∑

c ′∈C exp(v · f(x , y , c ′))

Where x is the input, y is the grounding, c is a correspondence variable, fis the array of features, and v is the array of feature weights.

[1] M. Collins, Log-Linear Models.http://www.cs.columbia.edu/~mcollins/loglinear.pdf

Page 10: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Log-linear Model – Training

Feature weights v are trained according to data from a corpus ofexamples.

The aim of training is to maximize the objective function:

Objective function and gradient

L′(v) =∑i

log p(ci | xi , yi ; v)− λ

2

∑k

vk2

(∇L′)(v)k =∑i

fk(xi , yi , ci )−∑i

∑c∈C

p(c | xi , yi ; v)fk(xi , yi , c)− λvk

The LBFGS optimization method [7] efficiently maximizes L′ whileconsuming little space.

[7] Byrd, R. H., Lu, P., Nocedal, J., Zhu, C. A Limited Memory Algorithm for BoundConstrained Optimization. 1995.

Page 11: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

The Problem

Number of possible individual groundings is O(n2) in the number ofobjects

Adding in sets of groundings makes it 2O(n2)

Page 12: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

The Problem

With 17 objects and 8 relations, the number of sets of constraints is

28×(17+8×17)2 = 3.08× 1056374

Page 13: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Partitioning Grounding Spaces

In many situations, most groundings are irrelevant

Partition the grounding space to eliminate irrelevant objects fromconsideration

Page 14: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Rules

Aim of rules is to partition grounding spaces to only include pertinentgroundings

Example

World: WorldObject(0, ‘robot’), WorldObject(1, ‘crate’),

WorldObject(2, ‘box’)

“approach the box” → {Rule(‘box’), Rule(‘robot’)}

Effectiveness of rules increases with complexity of environment andgrounding spaces

Page 15: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Hierarchical Grounding Graph

Run inference on space of rules

Apply result to grounding spaces in grounding graph model

Run inference in graphical model on partitioned grounding spaces forefficient grounding

Page 16: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Hierarchical Grounding Graph

true true true true true

approach the box near the chair

λ1 λ2 λ3 λ4 λ5

φ11 φ21 φ31 φ41 φ51

γ11

512γ21 γ31

1γ41

2γ51

1

f11 f21 f31 f41 f51

rules

true true true true true

φ12 φ22 φ32 φ42 φ52

γ12

800γ22 γ32

16γ42

272γ52

16

f12 f22 f32 f42 f52

Figure: Hierarchical Grounding Graph for “approach the box near the chair”

Page 17: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Score Evaluations

20 40 60 80 100 120 140 1602

2.63

3

4

5

6

Background objects

Sco

reev

alu

atio

ns

(log

10)

Score Evaluations for G3 Model and Hierarchical G3 Model

G3 model

Hierarchical G3 model

Page 18: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Run-time

20 40 60 80 100 120 140 1600

30

60

90

120

Background objects

Tim

e(s

)Runtime for G3 Model and Hierarchical G3 Model

G3 model

Hierarchical G3 model

Page 19: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Holodeck Experiment

Page 20: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Future Work

Expand space of rules to handle region and constraint types

Implement spatial features with regards to physical world model

Improve optimization routine (current runtime is impractical)

Test on Distributed Correspondence Graph model [3]

Handle parse ambiguity

Support more sophisticated sentence structures

Rigorous testing in more complex environments

Compute bounds on the efficiency of the algorithm

[3] T.M. Howard, S. Tellex, and N. Roy, A Natural Language Planner Interface forMobile Manipulators, to appear in the Proceedings of the 2014 International Conferenceon Robotics and Automation. 2014.

Page 21: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Acknowledgements

Thank you to

MIT PRIMES

Dr. Thomas Howard

Professor Nicholas Roy

Dr. Marec Doniec

Our parents

Page 22: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Bibliography I

[1] M. Collins, Log-Linear Models.http://www.cs.columbia.edu/~mcollins/loglinear.pdf

[2] S. Tellex, T. Kollar, S. Dickerson, M. Walter, A. Banerjee, S. Teller,and N. Roy, Approaching the Symbol Grounding Problem withProbabilistic Graphical Models. 2013.

[3] T.M. Howard, S. Tellex, and N. Roy, A Natural Language PlannerInterface for Mobile Manipulators, to appear in the Proceedings of the2014 International Conference on Robotics and Automation. 2014.

[4] J. Cocke, and J. Schwartz, Programming languages and theircompilers: Preliminary notes. 1970.

[5] D. Younger, Recognition and parsing of context-free languages in timen3. 1967.

[6] T. Kasami, An efficient recognition and syntax-analysis algorithm forcontext-free languages. 1965.

Page 23: Inferring the Structure of Probabilistic Graphical Models ...Inferring the Structure of Probabilistic Graphical Models for E cient Natural Language Understanding Istvan Chung Oron

Bibliography II

[7] Byrd, R. H., Lu, P., Nocedal, J., Zhu, C. A Limited Memory Algorithmfor Bound Constrained Optimization. 1995.


Recommended