+ All Categories
Home > Documents > Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

Date post: 18-Jan-2018
Category:
Upload: maximillian-fields
View: 218 times
Download: 0 times
Share this document with a friend
Description:
How can we see in depth? Problem: When light is focused on the retina to form an image, we go from 3 dimensions in the world to 2 dimensions on the image. We have lost the depth dimension. Question: How, then, can we see depth? Answers: Monocular cues to depth: from texture, motion, shading, etc. Binocular stereopsis: This uses the slightly different images on the two retinas to compute relative depth.
18
1 Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Transcript
Page 1: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

1

Perception and VR

MONT 104S, Fall 2008Lecture 8

Seeing Depth

Page 2: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

2

How can we see in depth?

Problem: When light is focused on the retina to form an image, we go from 3 dimensions in the world to 2 dimensions on the image. We have lost the depth dimension.

Question: How, then, can we see depth?

Answers:Monocular cues to depth: from texture, motion, shading, etc.

Binocular stereopsis: This uses the slightly different images on the two retinas to compute relative depth.

Page 3: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

3

Monocular cues to Depth When we view a scene with one eye, we still interpret the 3D scene fairly accurately.

The brain uses cues to the distance to estimate how far away things are. Some of these are listed here:

•Height in the visual field•Size of familiar objects (and relative object size)•Texture gradients•Linear perspective•Shading and shadows•Relative Motion (motion parallax)

Page 4: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

4

Height in the Visual FieldIn normal images, things that are farther away often have images that are higher in the visual field.

The visual system uses this height in the visual field as a cue to distance.

We then adjust our interpretation of the size of the object based on this distance.

(Note: familiar object size also gives a cue to depth here).

Page 5: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

5

Relative object size•Recall that with perspective projection, the image size of a nearby object is larger than for a distant object.

•Therefore, 2 similar objects of different sizes may be perceived at different distances.

•Demo: http://psych.hanover.edu/krantz/art/rel_size.html

•This cue works better when the objects are familiar and have known sizes (see previous slide).

Page 6: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

6

Texture Gradients•When we view a textured surface, the nearby texture is larger and coarser and the more distant texture is more fine grained. The texture forms a gradient from course to fine.•This is a consequence of perspective projection.

Demo: http://psych.hanover.edu/krantz/art/texture.html

Page 7: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

7

Linear PerspectiveBecause the size of images decreases with distance, the gap between 2 parallel lines will become smaller as the lines recede into the distance. This is known as linear perspective.

Our visual system uses this convergence of lines as a cue to distance.

Demo: http://psych.hanover.edu/krantz/art/linear.html

Page 8: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

8

Shape from ShadingShading in an image can give cues to the 3D shape of an object.The brain appears to assume the light source is above the scene.

In these images, when the shadow is on the bottom, we see bumps.When it is on the top, we see indentations.

Page 9: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

9

Motion ParallaxAs we learned in the motion lecture, for a moving observer, the images of nearby objects move faster than the images of distant objects.

The difference in image speed, known as motion parallax, can signal depth differences.

Demo: http://psych.hanover.edu/krantz/MotionParallax/MotionParallax.html

Structure from motion: Relative motion in the image can also be used to determine the 3D shape of objects.

Demo: http://www.viperlib.org/(General depth images, page 7, rotating cylinder).

Page 10: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

10

Moving ShadowsCombining motion and shadows can lead to some interesting perceptions of motion in depth.

Demo: http://www.viperlib.org/General Depth Images, page 1.

The only difference in these images is the position of the shadow.

Page 11: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

11

Binocular Stereo

When viewed with a red filter over the left eye and a cyan filter over the right eye, this image shows a 3 dimensional view of Mars.

The fact that our two eyes see the world from different positions allows the brain to use differences in the image to compute depth. This is known as stereopsis or stereo vision.

A 3D Mars Rover

Page 12: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

12

Binocular Stereo

•The image in each of our two eyes is slightly different.

•Images in the plane of fixation fall on corresponding locations on the retina.

•Images in front of the plane of fixation are shifted outward on each retina. They have crossed disparity.

•Images behind the plane of fixation are shifted inward on the retina. They have uncrossed disparity.

Page 13: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

13

Crossed and uncrossed disparity

crossed (positive) disparity

uncrossed (negative) disparity

plane of fixation

1

2

Disparity is the amount that the image locations of a given feature differ in the two eyes.

Page 14: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

14

Stereo processingTo determine depth from stereo disparity:

1) Extract the "features" from the left and right images

2) For each feature in the left image, find the corresponding feature in the right image.

3) Measure the disparity between the two images of the feature.

4) Use the disparity to compute the 3D location of the feature.

Page 15: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

15

Exam Review TopicsThere is a list of exam review topics at:http://mathcs.holycross.edu/~croyden/mont104S/notes/exam1_review.html

Expect short answer questions on any of the topics listed.

Examples:1) Why is vision considered an "ill-posed" problem? List 2

assumptions made by the visual system to deal with this problem.

2) What is the difference between rods and cones?

Page 16: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

16

More Sample Questions3. Describe the Craik-Obrien-Cornsweet illusion. What does it tell us about how the brain processes changes in contrast?

4. Diagram the neural circuit for direction selectivity. Briefly explain how it works.

5. Why is it that we can match any perceived color by a combination of red, blue and green lights?

6. List 3 monocular cues to depth. Briefly explain each one.

Page 17: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

17

Problem 1.Compute the 2D image positions for the following 3D points. Assume the image plane is 2 cm behind the pinhole (center of projection). All units are given in cm.

Point 1: (5, 10, 50)

Point 2: (16, 24, 8)

Point 3: (25, 45, 50)

Page 18: Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth

18

Problem 2.+6+6+6+6-2-2-2-2-2-2-2-2-2-2-2-2

The receptive field shown is partially in light and partially in dark. What is the change in firing rate of the neuron with this receptive field for this light stimulus? Assume that the partially covered squares are 1/2 covered.


Recommended