+ All Categories
Home > Documents > Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and...

Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and...

Date post: 11-Jan-2016
Category:
Upload: kellie-parks
View: 215 times
Download: 0 times
Share this document with a friend
Popular Tags:
41
Robots and AI
Transcript
Page 1: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Robots and AI

Page 2: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Marr’s 3-level of vision: 1) Computational or task analysis2) Representation and algorithm3) Implementation = Neuro-level

→ Computational level of cognition – interconnected (?) to neuronal level

↔ Relationship perception-action-brain/mind - strong interconnected

• Evolutionary framework

Page 3: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Classical AI • Abstract (physical irrelevant) • Individual (mind = locus of intelligence)• Rational (reasoning → Intelligence) • Detached (thinking separated from

perception + action) (Smith 99 in Ekbia 08)

• Classical approach = “Sense-think-act” Percep-al mechanism → 3D visual scene = Input to reasoning/planning centres → Calculate the action + commands to motor → Action

Page 4: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

vs.

• “Interactive vision” (Churchland, Ramachandran, Sejnowsky 94):

Low-level perception involves motor routines

• Real-world actions → Computations

• Rs = Not passive information but “direct recipe for action” (Clark 01)

Page 5: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Early Robots - Navigating with Maps

• Social insects: communication (honeybees)

SHRDLU

• Simulated robot (MIT) operated in a simulated blocks microworld

• Graphic interface and a video screen that displayed its operations in a visual manner

• Written language interface - followed commands given in ordinary English + answered questions about its “motivations” for doing things in sequence

Page 6: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Asked why it picked up little pyramid among its blocks: “To clear off the red cube.”

• Able to remember and report its previous actions (touching a particular pyramid before putting a particular block on a particular small cube) (Crevier 1993 in Ekbea 2008)

Page 7: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

New Robots: Navigating Without Maps• Toto = Robotic cat - navigates corridors of office

without a priori map. Uses compass - keeps track of its ordered interactions with various landmarks (a wall on right, a corridor). Landmarks - used to construct a map = Not explicit but part of robot (Mataric 92)

• Luc Steels - Simulated society of robots …, self-organize into a path and attract other robots. Descriptions of paths = Not in robot (Steels 90)

• Genghis = Robotic insect - walks toward anymoving source of infrared radiation, steers to keep its target in sight, scrambles over obstacles in its way, no internal notion of “toward/ahead/over” (Brooks 02 in Ekbea 08)

Page 8: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

The robot Genghis.

Page 9: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Brooks ‘97: “The world is its own best model”1. Situatedness: Embedded in world, NOT deal

with abstract descriptions (logical sentences, plans) vs. its sensors - “here and now” of world → Behavior

2. Embodiment: Physical body + Experiences world

3. Intelligence: “Intelligence - determined by dynamics of interactions with world”; Evolution: AI - “low-level”!

4. Emergence: Complex behavior - emerge - interactions among primitive tasks/modules;

“Intelligence is in eye of observer.”

Page 10: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Brooks (91)

• Disembodied programs for reasoning and inference in abstracted natural language processing, visual scene analysis, logical problem solving = Mistake

vs

• Embedded in dynamic realworld situations, integrating perception and action in real time → Fluid embodied adaptive behavior (Wheeler 05)

• Insects → Humans

Page 11: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Rodney Brooks (1991) “new robotics” • “Subsumption architecture”: Robot - 3 layers • Each layer - a function, input to motor action ↔ Separate control system (a layer = hard-wire

finite state machine) for each task • 3 layers: avoiding obstacles, moving randomly,

moving toward a location• Coordination between layers (external input -

one device turns off another turns on) → Sequences of a serial processes

• Subsumption architecture = Decomposition of activities horizontally by task, not vertically by function↔NO central processor/R/modules

Page 12: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Robot Herbert (Connell 1989)• Collect soft drink cans on tables• “Sense-think-act” view vs. Collection of sensors +

Independent behavioural routines (ring of ultrasonic sound sensors, robot halts in front of object)

• Difference: Random movement - interrupted if its visual system detects a “table-like outline” → New function: Sweeping surface of table

• If detected → “Robot rotates until can is centred in its field of vision”

• Arm - touch sensors skim table surface until a can encountered, grasped, collected

• Movement → Perception = Not passive phenomena• Perception and action - Strong interconnected

(Clark 2001)

Page 13: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• ”Mirror neurons”: Neurons in monkey - action oriented, context dependent- implicated in both self-initiated activity + passive perception.” (Di Pellegrino, all 92)

• Neurons - activated monkey observes + performs an action

Perception,action,cognition - interconnected

• Evolution line: Brain = “organ of environmentally situated control”

Page 14: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Clark (2008)

• Honda’s Asimo – Most advanced humanoid robot = “Passive-dynamic walker”

vs.

• Active robot = Environment is “incorporated” in robot’s functions

• Pfeifer et al. (2006) - “Ecological control”:

“Part of ‘processing’ - by dynamics of agent-environment interaction, and only sparse neural control needs to be exerted when self-regulating and stabilizing properties of natural dynamics can be exploited.”

Page 15: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Active robots • Kuniyoshi et al. 2004: “Rolling and rising” motion• Iida and Pfeifer’s (2004): Running robot Puppy• Pfeifer and Bongard (2007) → Clark - Principle of

Ecological Balance:• Task environment - match between agent’s sensory,

motor, neural systems + task-distribution betw. morphology, materials, control, environment

• “Matching” → Responsibility for adaptive response “not all processing is performed by brain, but by morphology, materials, environment → ‘Balance’/task-distribution between different aspects of an embodied agent” (Pfeifer et al. 2006)

Page 16: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Toddler robot

• “Can learn to change speeds, go forward and backward, and adapt on different terrains, including bricks, wooden tiles…

• Similar to a child - learns to “complex evolved morphology and passive dynamics of its own body”

• Can exploit passive dynamics of its own body for controlling its movements

(Not for passive robot)

Page 17: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Fitzpatrick et al. (2003) - BABYBOT platform: Information about object boundaries is furnished by “active object manipulation” (“pushing + touching objects in view”)

• “Learns about boundaries by poking + shoving”; Uses motion detection to see its own hand–arm moving

• The infants “grasping, poking, pulling, sucking, and shoving create a flow of time-locked multimodal sensory stimulation.”

Page 18: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Multimodal input stream aid category learning and concept formation! (Lungarella, Sporns, and Kuniyoshi 2008; Lungarella + S 2005)

• “Self-generated motor activity” = Complement to “neural information-processing” → “Information structuring” by motor activity and “information processing” by neural system =

Continuously from embodiment to cognitive extension linked to each other through sensorimotor loops.” (Lungarella + S 05)

Page 19: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

COG (MIT, Brooks’ team)• An upper-torso humanoid body that learns

to interact with people through “senses”

Characteristics: • Embodied–body/parts similar human body• Embedded – it is “socialized” (minor)• Developing – “baby” version Kismet• Integrated – equipped with + to integrate

data from equivalents of various sensory organs (Ekbia 2008)

Page 20: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

COG (MIT)

Page 21: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• “The distinction between us and robots is going to disappear.” (Brooks 2002)

Page 22: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

COG

• … cross-modal binding of incoming signals - display common rhythmic signatures → Robot in learning about objects + its own body

• Detects rhythmic patterns in sight, hearing …

• Deploys a binding algorithm to associate signals that display same periodicity

• Bindings → COG learn its own body parts by binding visual, auditory, proprioceptive signals (Fitzpatrick, Arsenio 04)

Page 23: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Cog group • From natural selection to child development• “Adult robots” from “baby robots”

Kismet: • Social interaction robots-humans• Eyebrows (each onewith two degrees of

freedom: lift and arc) • Ears (2 degrees of freedom: lift and rotate)• Eyelids + mouth (1 degree: open/close)• Two microcontrollers (driving robot’s facial

motors + “motivational system” (Ekbia 2008)

Page 24: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Kismet: emotive facial expressions indicative of anger, fatigue, fear, disgust, excitement, happiness, interest, sadness, surprise

Page 25: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Perception + action = Separate processes mediated by a “brain”/central processor

vs.

• Situated approach: Perception+action = Essential (Central processing + Rs of world – not important) (Ekbea 2008)

• There are Rs/accumulations of state, but refer only to internal workings of system; meaningless without interaction with outside world. (Brooks 1998 in Ekbea)

Page 26: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• “Eye contact” with human being → Social interaction; Imitate the head-nodding

• Progressive development (imitation + joint attention)

• Decomposed them 4 stages:- Eye contact - Gaze-following - Imperative pointing (trying to obtain an

object that is out of reach by pointing at it - Declarative pointing (drawing attention to a

distant object by pointing at it) (p. 270) (Ekbia 2008)

Page 27: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Developmental psychology:

• Purely mentalistic (it explains behaviors in terms of internal Rs)

vs.

• Behaviorist (behaviors = NO Rs)

• Both

• Robots = Testbed for evaluating predictive power + validity of psychological

theories through repeatable experiments

• Kismet - Not what children do; but how they do it

Page 28: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Robot - learn using an “open-ended variety of internal, bodily, or external sources of order.” = “Natural-born cyborgs” (Clark 03 in Clark 08) → Body=“key player on problem-solving”

• New trend in cognitive science: “Loosing bonds between perception and action”!

• Hybrid model = Relating sensorimotor information with cognition

• “Inner and outer elements (distributed problem-solving ensemble) must interact = Integrated cognitive whole” (Clark 08)

Page 29: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• David Kirsh (1991) Today earwig, tomorrow human

• 97 % of human activity = Concept-free = No concepts from traditional AI (Kirsh 91)

Page 30: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Wheeler: 2 Cartesian dogmas = Distinctions(1) Mind-world (2) Mind-body

• Rejects R + computation• Primary function internal processes = For

sensations + control action + basic sensoriomotor processes - not isolated higher processes = Heideggerian paradigm

(Husserl - phenomenology, Heidegger, Merleau-Ponty, Dreyfus, etc.)

Page 31: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Anti-representationalism - “2 treats to R”:• 2 “treats” against explanation of online

behaviour needs “R”:

(1) If extra-factors are necessary to explain the behaviour of a system (“non-trivial causal spread”) → No R

(2) R view = “Homuncularity” - rejected: causal contribution of each component of a system is context-sensitivity and variable over time (“continuous reciprocal causation”)

• Ex-s: Brooks (1991) + Franceschini et al. (1992) with a robot with elementary motion detectors avoiding obstacles

Page 32: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Clark and Wheeler: Causal spread (1999) = Internal elements depend upon certain causal factors external to system

• Ex: Computational neuroethology of robots (Dave Cliff, Cliff, Harvey and Husbads)

• Simulation of robot+ room - evolved to control robot moving in rooms

• Online-offline cognition blurred if we reject arbitrariness (different classes for same function) and homuncularity

Page 33: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Wheeler (‘05): Homuncularity → Modularity

• Continuous reciprocal causation - multiple interactions + dynamic feedback loops

(i) Causal contribution of each component in system determines + determined by causal contributions of large numbers of other components in system

(ii)Contributions change radically over time

→ Dynamical holistic perspective (against modularity R)

Page 34: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Mirrors work of nature in creating man + consciousness

• Clark (01): Robotics (AI) - Low-level systems - strong relation body- action-environm.→ Adaptive behavior

Involve emergence and collective effects

vs.

Classical model (“hear-localize-locomote” routine = Task decomposition + identifies a sequence of subtask)

Page 35: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Webb’s cricket phonotaxis

• Male cricket’s song has to be hear

• Identify, localize by female that has to locomote toward it

• Cricket anatomy + neurophysio. (ears + tracheal tube)

• “Vibration - greater at ear nearest to the sound source → Orientation and locomotion” (Clark)

Page 36: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Cricket's tracheal tube transmits sounds of desired calling song frequency- phase shifts - Particular wavelength

Thus:

• No general mechanism for identifying direction sounds

• No actively discriminate song of its own species from other songs

• Other sounds - structurally no generating response

Page 37: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• No - general purpose capacities (pattern recognition + sound localization) to mate detection

• It exploits highly efficient but (because) special-purpose strategies

• No model of its envir. + apply logico-deductive inference --- action plans

• No central sensory information store for integrating multimodel inputs

• No Rs - not necessary symbolic interpretation to explain how system functions

Page 38: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Harold Cohen’s Aaron, a computer-artist whose products have enjoyed approval of art world, being exhibited in respected museums around world (Ebkea p. 291-4)

• An autonomous robot embodies principles of goal-seeking and scanning that characterize animal behavior

Page 39: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

General ideas• A-life = GA, CA, networks controller robots

• The pair artificial life-biology in parallel with AI-psychology

• Langton: Synthetic strategy → “A-life” - synthetic approach for understanding evolution + operation living systems

→ Build simulated systems from components: what emerges

Page 40: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

• Relationship life-mind reflects life-artificial

→ Definition of Life = Obscure

Life – Properties:

• Autopoesis

• Autocatalysis elements

• Self-reproduction

• Genetics and metabolism

• Cluster concept – multiple features

Page 41: Robots and AI. Marr’s 3-level of vision: 1)Computational or task analysis 2)Representation and algorithm 3)Implementation = Neuro-level → Computational.

Cellular Automata

• http://mathworld.wolfram.com/CellularAutomaton.html


Recommended