Post on 13-Jan-2016
description
transcript
NLog-like Inference and Commonsense Reasoning
Len Schubert
University of Rochester
Student participants: Ben Van Durme, Ting Qian, Jonathan Gordon, Karl Stratos, Adina Rubinoff Support: NSF (Grants IIS-1016735 and IIS-0916599), ONR STTR N00014-10-M-0297
EL & EPILOG: Representation & inference for NLU, common sense
(L. Schubert, C-H Hwang, S. Schaeffer, F. Morbini, et al., 1990 – present)
2
episode
set
hier2
colormetanumberstring
time
typeequalityparts
other
LOGICAL INPUT
LOGICAL OUTPUT
EPILOG core
SpecialistInterface
“A car crashed into a tree. …” (some e: [e before Now34]
(some x: [x car] (some y: [y tree] [[x crash-into y] ** e])))
“The driver of x may
be hurt or killed”
Episodic Logic (EL):
A Montague-inspired, event-oriented extensionOf FOL, with NL-like expressive devices.
3
THE EPISODIC LOGIC/EPILOG PERSPECTIVE,
Reasons for hypothesizing a language-like internal representation:
• Anthropology, cognitive science: concurrent appearance of thinking, language
• Simplicity of assuming NL “Mentalese”
• All our symbolic representations, from logic to programming languages to semantic nets, etc., are derivative from language
• Can one seriously believe that its just a coincidence that entailment can be understood in terms semantic entities corresponding 1-1 with syntactic phrases (Montague, categorial grammar)??
• Recent progress in applying “natural logic” to inferring entailment relations.
Guided thedevelopmentof EPISODICLOGIC
4
Universal semantic resources of natural languages
• Ways of naming things
• And/or/not/if-then/… • Every/some/no/ …• Ways of ascribing properties and relations to entities
BUT THAT’S NOT ALL!
• Generalized quantifiers (Most women who smoke)
• Intensionality (is planning a heist; resembles a Wookiee)
• Event reference (Everyone asked questions; THAT prolonged the meeting)
• Modification of predicates and sentences (barely alive, dances gracefully, Perhaps it will rain)
• Reification of predicates and sentences (Xeroxing money is illegal; That there is water on the Moon is surprising)
• Uncertainty (It will probably rain tomorrow; The more you smoke, the greater your risk of developing lung cancer)
•
• Quotation and meta-knowledge (Say “cheese”; How much do you know about description logics?)
directlyenabled in EL
So, at least FOL!
5
Episodic Logic (EL) examples
• Restricted quantifiers “Most laptops are PCs or MACs”
(Most x: [x laptop] [[x PC] or [x MAC]])
• Event relations “If a car hits a tree, the driver is often hurt or killed”
(Many-cases e: (some x: [x car] (some y: [y tree] [[x hits y] ** e])) [[[(driver-of x) (pasv hurt)] or [driver-of x) (pasv kill)]] @ e ])• Modification and reification
“He firmly maintains that aardvarks are nearly extinct”
(Some e: [e at-about Now17] [[He (firmly (maintain (that [(K (plur aardvark)) (nearly extinct)])))] **
e])
Note: Predicates are infixed
Representing Meta/Self-Knowledge in EL: Schemas (substitutional quantification) + quasi-
quotes
“I know the names of all CSC faculty members”
6
( x: [x member-of CSC-faculty] ( y: [‘y name-of x] [ME know (that [‘y name-of x])]))
AA
subst
There is no CSC faculty member whose name I knowto be ‘Alan Turing’.
Therefore there is no faculty member whose name is‘Alan Turing’.
• Can replace phrases by more general [more specific] ones in positive- [negative-] polarity environments;
e.g., Several trucks are on their way Several vehicles are on their way;
If a vehicle is on its way, turn it back If a truck is on its way, turn it back
• Exploiting implicatives/factives,
e.g., X manages to do Y X do Y;
X doesn't manage to do Y ~> X doesn't do Y;
X knows that Y Y;
X doesn’t know that Y Y;
• Full disambiguation not required; e.g., “several”, “on their way” can
remain vague and ambiguous without disabling the above inferences
Ideas behind Natural Logic (Nlog) (van Benthem, van Eijck, Sanchez Valencia, Nairn, Condoravdi,
Karttunen, MacCartney & Manning, etc.)
• EPILOG inference is in essence polarity-based: replacing subformulas by consequences/anti-consequences in +ve/-ve environments (plus natural deduction rules, specialists)
• The equivalent of Nlog inference are readily encoded as axioms and rules in EPILOG 2. E.g., we have duplicated MacCartney & Manning’s illustrative example,
Jimmy Dean refused to move without his jeans
James Dean didn’t dance without pants,
but also examples requiring background knowledge (beyond natural logic).
• Details & examples to follow.
NLog-like inference in EPILOG 2
Examples of implicative axioms (all_pred p (all x ((x dare (ka p)) => (x p))))),
(all_pred p (all x ((not (x dare (ka p))) => (not (x p))))))
Similarly for other implicatives; also attitudes (stylized rules):
X decline to P => X not P
X not decline to P => (probably) X P
X agrees to P => X P
X does not agree to P => (probably) not X P
X doubts that W => X believes probably not W.
Example of inference rules for a factive verb:
(all_wff w (all x ((x know (that w)) ---> w)))),
(all_wff w (all x ((not (x know (that w))) ---> w))))
Headline examples (by Karl Stratos)• Vatican refused to engage with child sex abuse inquiry (The Guardian: Dec 11, 2010).
• A homeless Irish man was forced to eat part of his ear (The Huffington Post: Feb 18, 2011).
• Oprah is shocked that President Obama gets no respect (Fox News: Feb 15, 2011).
• Meza Lopez confessed to dissolving 300 bodies in acid (Examiner: Feb 22, 2011)
In EPILOG (neglecting tense): (s '(Vatican refuse (ka (engage-with Child-sex-abuse-
inquiry)))) (s '(some x (x (attr homeless (attr Irish man))) (x (pasv force) (ka (l y (some r (r ear-of y) (some s (s part-of r) (y eat
s)))))))) (s '(Oprah (pasv shock) (that (not (Obama get (k respect)))))) (s '(Meza-Lopez confess (ka (l x (some y (y ((num 300) (plur body))) (x dissolve
y))))).
Inferred in fractions of a second (& returned in English):
The Vatican did not engage with child sex abuse inquiry. An Irish man did eat part of his ear, President Obama gets no respect, and Meza Lopez dissolved 300 bodies in acid.
Larger-scale factive/implicative/attitudinal inferences in EPILOG 2
Karl Stratos has used his axiomatic factivity/ implicativity lexicon on100+ EPILOG-encoded Brown corpus examples; e.g.,
e.g., I know that you wrote this in hurry.
You wrote this in hurry.
e.g., They say that our steeple is 162f high
Probably they believe that our steeple is 162f high
Evaluation: 108 sentences from Brown corpus, 141 inferences…
92% were rated as good (75%) or fairly good (17%)
(5 judges)
• Lexical axiom: (all_pred p (all x (all y (all e1 ((x ask-of.v y (Ka p)) ** e1)
((x convey-info-to.v y
(that ((x want-tbt.v
(that (some e2 (e2 right-after.p e1)
((y p) ** e2)))) @ e1))) * e1)))))
• Given: John asked Mary to sing ((John.name ask-of.v Mary.name (Ka sing.v)) ** E1)
• Question: Did John convey to Mary that he wanted her to sing?
((John.name convey-info-to.v Mary.name
(that ((John.name want-tbt.v
(that (some e2 (e2 right-after.p E1)
((Mary.name sing.v) ** e2)))) @ E1))) * E1)
Answered with YES in .001 sec
Pushing the limits of NLog --E.g., entailments of asking someone to do
something
Simple inference beyond the scope of NLog (Allen's Monroe domain): Every available crane can be used to hoist rubble onto a truck. The small crane, which is on Clinton Ave, is not in use. Therefore, the small crane can be used to hoist rubble from the collapsed building on Penfield Rd onto a truck.
• Every available crane can be used to hoist rubble onto a truck (s '(all x (x ((attr available) crane)) (all r (r rubble) ((that (some y (y person) (some z (z truck) (y (adv-a (for-purpose (Ka (adv-a (onto z) (hoist r)))) (use x)))))) possible))))
• The small crane, on Clinton Ave., is not in use. (s '(the x (x ((attr small) crane)) ((x on Clinton-Ave) and (not (x in-use)))))
• Every crane is a device (s '(all x (x crane) (x device)))
• Every device that is not in use is available (s ‘(all x ((x device) and (not (x in-use))) (x available)))
• Can the small crane be used to hoist rubble from the collapsed building on Penfield Rd onto a truck? (Answered affirmatively by EPILOG in .127 sec) (q (p ‘(the x (x ((attr small) crane)) (some r ((r rubble) and (the s ((s (attr collapsed building)) and (s on Penfield-Rd)) (r from s))) ((that (some y (y person) (some z (z truck) (y (adv-a (for-purpose (Ka (adv-a (onto z) (hoist r)))) (use x)))))) possible)))))
An example requiring still more world knowledge Most of the heavy resources are in Monroe-east. Therefore: - Few of the heavy resources are in Monroe-west; - Not all of the resources are in Monroe-west
Some general knowledge:
• If most P are not Q then few P are Q: (s '(all_pred P (all_pred Q
((most x (x P) (not (x Q))) -> (few x (x P) (x Q))))))
• “Heavy” in premodifying position is subsective (s '(all_pred P (all x (x ((attr heavy) P)) (x P))))
• “If most P are Q, then some P are Q (existential import of “most”)
(s '(all_pred P (all_pred Q ((most x (x P) (x Q)) -> (some x (x P) (x Q))))))
• All Monroe resources are in Monroe. A thing is in Monroe iff it is in Monroe-east or Monroe-west; and iff it is in Monroe-north or Monroe-south; nothing is in both Monroe-east and Monroe-west; or in both Monroe-north and Monroe-south:
(s '(all x (x Monroe-resources) (x loc-in Monroe)))
(s '(all x ((x loc-in Monroe) iff
((x loc-in Monroe-east) or (x loc-in Monroe-west)))))
(s '(all x ((x loc-in Monroe) iff
((x loc-in Monroe-north) or (x loc-in Monroe-south)))))
(s '(all x ((not (x loc-in Monroe-east)) or (not (x loc-in Monroe-west)))))
(s '(all x ((not (x loc-in Monroe-north)) or (not (x loc-in Monroe-south)))))
(Example requiring still more world knowledge, cont’d)
• There are some heavy Monroe resources.
Most of the heavy Monroe resources are located in Monroe-east (s '(some x (x ((attr heavy) Monroe-resources))))
(s '(most x (x ((attr heavy) Monroe-resources)) (x loc-in Monroe-east)))
Questions:
• Are few heavy resources in Monroe-west? (q (p '(few x (x ((attr heavy) Monroe-resources)) (x loc-in Monroe-west))))
Answer is “yes”.
• Are all Monroe resources in Monroe-west? (q (p '(all x (x Monroe-resources) (x loc-in Monroe-west))))
Answer is “no”, because: Most heavy resources, hence some heavy resources, hence some resources, are in Monroe-east; but whatever is in Monroe-east is not in Monroe-west, hence not all resources are in Monroe-west.
• Lexical knowledge (for Nlog-like & other inference)
• Semantic patterns (as initial, underspecified world knowledge and for
parsing/interpretation)• World knowledge (for more general
reasoning)• Mapping Treebank parses into EL (for NL-
based inference)
Trying to Scale up Knowledge,
and mapping NL into EL
• Entailment, synonymy, and exclusion relations among lexical items, by starting with distributional similarity clusters, and training a classifier to select the correct relation; initial results ~80% accurate
• Knowledge engineering of a large collection of factive, antifactive, and implicative verbal predicates for use in EPILOG, gleaned from various sources and expanded via VerbNet, etc. (undergrad Karl Stratos has been the mainstay of this effort); 250 lexical items with their semantic “signatures”
• preliminary set of detailed, event-oriented lexical axioms, leveraging Palmer's VerbNet (VN); (Adina Rubinoff); three stages:
- axiomatized ~100 semantic “primitives” (MOVE, SEE, LEARN, MAKE, …)
- creating an axiom schema for each VN class, in terms of “primitives” and “(predicate) parameters”
- providing parameters for the verbs in each class (e.g., the states resulting from break, repair, melt, etc., some inferrable from VN)
Lexical Knowledge Acquisition
KNEXT:
KnowledgeExtractionFrom Text
A PERSON MAY BUY FOOD;A HOUSE MAY HAVE WINDOWS;A COMEDY MAY BE DELIGHTFUL;A BEHAVIOR CAN BE STRANGE;LEISURE MAY BE DEVOTED TO PLAY;
…
The KNEXT project: General KNowledge EXTraction from text
(L. Schubert, M. Tong, J. Sinapov, B. Van Durme, T. Qian, J. Gordon, …)
18
A starting point for world knowledge acquisition:
General “factoids”, orsemantic patterns
The KNEXT system: Functional architecture
compute LFs
80 regularphrase patterns,paired with sem- antic rules
extract & abstract propositions
verbalize and filter propositions
adjust phraseStructure forInterpretation
identify temporalphrases, etc.
proper namegazetteer; “of”- knowledge, etc.
sentence & phrase structure
adjusted input
sets of LFs
propositional LFs
abstract LFs andEnglish output
[S [NP I] [VP had [NP a terrible flu] [NP last year]]]
[S [NP I] [VP had [NP a terrible flu] [NPtime last year]]]
[mePron haveV fluN], <a{n} x[x (attr terribleA fluN)]>
[<a{n} personN> haveV fluN], [<a{n} fluN> terribleA]
[<a{n} personN> haveV fluN], [<a{n} fluN> terribleA]
A PERSON MAY HAVE A FLU A FLU CAN BE TERRIBLE
“shallow”knowledge
Text corpora used, & example output…
20
Brown Corpus: 1 million words, with phrase structure ----> 117,000 factoids
British National Corpus: 100 million words, analyzed with Collins parser ----> several million factoids
Weblogs, Wikipedia (Jonathan Gordon): billions of words ----> 200 million factoids
Selected Brown examples: A PERSON MAY BELIEVE A PROPOSITION BILLS MAY BE APPROVED BY COMMITTEES A US STATE MAY HAVE HIGH SCHOOLS CHILDREN MAY LIVE WITH RELATIVES A COMEDY MAY BE DELIGHTFUL A BOOK MAY BE WRITE-ED (i.e., written) BY AN AGENT A FEMALE-INDIVIDUAL MAY HAVE A SPOUSE AN ARTERY CAN BE THICKENED A HOUSE MAY HAVE WINDOWS PROTESTS CAN BE ADAMANT A MALE-INDIVIDUAL MAY LEAD A FIGHT A TEAM CAN BE WINLESS LEGS MAY TWITCH INDIVIDUALS MAY SHARE A BED REVELATIONS MAY EMBARRASS TOWN OFFICIALS A BRICK FAÇADE MAY BE SHEARED OFF BY A SHOCK OF A QUAKE A TV-NETWORK MAY HAVE A SPOKESMAN A BARREL MAY CONTAIN HEATING OIL A LANGUAGE MAY BE MELLIFLUOUS
21
Abstracting from, and disambiguating, factoids
(Van Durme, Michalak & Schubert EACL’09)
“A CHILD MAY WRITE A LETTER”
“A JOURNALIST MAY WRITE AN ARTICLE”
TEXT
DOCUMENT
WRITTEN MATTER
PIECE OF WRITING
WRITTEN COMMUNICATION
COMMUNICATION
REPRESENTATION
CREATION (phys)
ARTICLE2 (e.g., clothing)
ARTICLE (legal) ARTICLE4
(grammar)
NONFICTIONPROSE
PROSE
LITERARYGENRE
EXPRESSIVE STYLE
ARTIFACT
WHOLE
PHYSICAL OBJECT
PHYSICAL ENTITY
ABSTRACT ENTITY
ENTITY
COMPOSITION (phys)
GENERALLY, IF X WRITES Y, Y IS A COMMUNICATION
WordNetontology
LETTER1
(missive) LETTER2
(alphabet) LETTER3
(landlord) LETTER5
(varsity)
LETTER4
(of the law)
ARTICLE1
(literary)
• Engineered rules have transformed tens of thousands of text-derived "possibilistic" factoids (such as that A TREE MAY HAVE A BRANCH, or A PERSON MAY EAT A SANDWICH) into "sharper" quantified formulas such as
(most-or-all x: [x tree]
(some y: [y branch] [x has-as-part y]))
(many x: [x person]
(at-least-occasional e
(some y: [y sandwich] [[x eat y] ** e]))),
i.e., most or all trees have at least one branch, and many people eat a sandwich at least occasionally.
• 1.5 million sharpened factoids have been obtained (accessible at http://www.cs.rochester.edu/research/knext/browse/); for 435 sampled sharpened factoids, about 60% were judged reasonable if based on reasonable unsharpened factoids (o/w about 40%).
Obtaining inference-capable knowledge by “sharpening” factoids
(J. Gordon & L. Schubert KCS’10)
• Use Tgrep on parsed sentences to find patterns such as
NP VP but didn’t VP , NP VP, expecting to VP NP BE ADJP {but|yet} ADJP,
i.e., where an expectation is implied (and perhaps denied). Apply rules to them that create slightly simplified / abstracted conditional statements, expressed as parse trees (not yet
LFs)• E.g., He stood before her in the doorway, evidently expecting to be
invited in If a male stands before a female in the doorway,
then he may expect to be invited in. Other sample rules: If a person texts a male, then he-or-she may get a
reply; If a pain is great, then it may not be manageable; If a person doesn’t like some particular store, then he-or-she may not keep going to it. • About 1 out of 200 sentences yields a rule (that survives
filtering); e.g., 29,000 rules from a 5.5 million sentence story corpus; of these more than 2/3 are judged to be reasonable.
Discovering commonsense entailment rules based on discourse cues
(J. Gordon & L. Schubert, TextInfer ‘11)
• Even lexical glosses are hard to interpret; e.g., (WordNet)
dance (V): move in a pattern, usually to musical accompaniment
What does “in a pattern” mean? (Cf. “move into / inside a pattern”) What does “to musical accompaniment” mean? (towards?)
• Open Mind factoids leave much unsaid; e.g.,
Something you might do while driving a car is crash
Who / what is crashing (into what)?
• Some (simple English) Wikipedia items are simple, clear, and complete; others, not so much … A car (also called an automobile) is a vehicle used to transport passengers. Cars usually have four wheels and an internal combustion engine.
Dance is when people move to a musical rhythm….
What does “is when” mean? (Cf. “Monday is when I go home”) What does “to a musical rhythm” mean? (Towards it? And does a marching band dance?)
What about direct interpretation of general statements (lexicon glosses,
Open Mind, Wikipedia, …)?
Computing initial logical forms in ELSome time-worn examples: “Time flies like an arrow”
[(K timeN) (adv-m (likeP <a{n} arrowN>) <pres flyV>)],
[(K (plur (nn timeN flyN))) <pres likeV> <a{n} arrowN>],
… + readings with timeV
“I saw the man with binoculars”
[IPron (adv-m (withP binocularsN) (<past seeV> <the manN>))],
[IPron <past seeV> <the (x [[x manN] & [x withP binocularsN ]])>))],
… + readings with <pres sawV>
25
“Shallow knowledge” (semantic patterns)…Time may fly; an arrow may fly;Seeing may be done with a viewing instrument
…should help with “gross ambiguities”!
Further disambiguating and elaborating
the initial LF
26
…all depend profoundly on lexical & world knowledge, and context
Finding referents of pronouns and other terms …“He tried to steal Donald Trump’s identity but couldn’t pull it off”
Scoping quantifiers …“Every man admires a certain woman” (his mother? Rosa Parks?)
Recovering “missing arguments” & comparison classes…“Some carbon monoxide leaked into the car, but its concentration was too low to pose a serious hazard
Expanding metonymy …“THIS LANE MUST EXIT” (vehicles travelling in this lane …)
Inferring temporal, causal, & other coherence relations…“I told Rocky he was a wimp. When I regained consciousness, …”
Inferring speaker/author intent…“Sir, you’re sitting in my seat”
Determining what is presupposed, and what is new…“Cro-Magnons usually roasted meat on a spit over a fire” (i.e., usually when preparing to eat meat!)
.
.
.
.
.
.
.
DiscussionThe most important remaining problems are - KB build-up - reliable mapping from English to a structurally
unambiguous, deindexed, reference-resolved EL form.
Does Nlog escape these problems? Not really:- A large KB is essential in either case- We need to generate inferences, not just verify them.
This cannot be done by alignment + word-level editing
- We need deindexed representations for general inference. If “I will soon stop talking”, were true in perpetuity, I would never stop – nor would anyone else using the pronoun “I”!
- Ambiguity/vagueness can be tolerated only to a limited extent, even in Nlog; “John had gerbils as a child” should not be regarded as entailing that John consumed, or gave birth to, small rodents as a child.
- If we actually want to understand language, we need to let world knowledge, not only lexical knowledge, play its role
Conclusions
28
The representational and inferential style of EL / EPILOG is close to that of Nlog;
EL / EPILOG also allow for more complex inferencesfrom lexical and world knowledge;
Ambiguity resolution and knowledge accumulation remain issues for both EL / EPILOG and Nlog.
.
.
.
• Schubert, Van Durme, & Bazrafshan, “Entailment inference in a natural- logic-like general reasoner”, AAAI Fall Symp. On Commonsense Knowledge (CSK’10), November 2010;
• Stratos, Schubert, and Gordon, “Episodic Logic: Natural Logic + Reasoning”, to appear.
• Gordon and Schubert, “Quantificational Sharpening of commonsense knowledge”, AAAI Fall Symp. On Commonsense Knowledge CSK’10), November 2010;
• Gordon and Schubert, “Discovering commonsense entailment rules implicit in sentences”, TextInfer 2011.
References