+ All Categories
Home > Documents > Binocular Helmholtz Stereopsis

Binocular Helmholtz Stereopsis

Date post: 16-Mar-2022
Category:
Upload: others
View: 3 times
Download: 0 times
Share this document with a friend
7
Binocular Helmholtz Stereopsis Todd E. Zickler Jeffrey Ho David J. Kriegman Jean Ponce Peter N. Belhumeur [email protected] [email protected] [email protected] [email protected] [email protected] Electrical Engineering, Yale University, New Haven, CT 06511 Computer Science and Engineering, University of California at San Diego, La Jolla, CA 92093 Computer Science, University of Illinois at Urbana-Champaign, Urbana, IL 61801 Computer Science, Columbia University, New York, NY 10027 Abstract Helmholtz stereopsis has been introduced recently as a sur- face reconstruction technique that does not assume a model of surface reflectance. In the reported formulation, corre- spondence was established using a rank constraint, neces- sitating at least three viewpoints and three pairs of images. Here, it is revealed that the fundamental Helmholtz stereop- sis constraint defines a nonlinear partial differential equa- tion, which can be solved using only two images. It is shown that, unlike conventional stereo, binocular Helmholtz stere- opsis is able to establish correspondence (and thereby re- cover surface depth) for objects having an arbitrary and unknown BRDF and in textureless regions (i.e., regions of constant or slowly varying BRDF). An implementation and experimental results validate the method for specular sur- faces with and without texture. 1 Introduction Helmholtz stereopsis has recently been introduced as a multi-view technique for estimating surface shape without requiring an assumed model of reflectance. Thus, unlike most existing methods, it enables the dense reconstruc- tion of scenes that contain surfaces with unknown, spatially varying, and arbitrary reflectance functions (BRDFs). In Helmholtz stereopsis, this is accomplished through the use of images that are collected in reciprocal pairs [21]. The first image in a reciprocal pair is acquired under a single point light source, and the second image is acquired after in- terchanging the camera and light source positions. Figure 1 shows a reciprocal pair, and Fig. 2 illustrates the acquisi- tion geometry. Reciprocal pairs have the unique property that the relation between intensities at corresponding image points depends only on surface shape and is independent of reflectance. This property follows directly from the symme- Figure 1. Two rectified images of a painted, plastic mannequin head acquired as a Helmholtz reciprocal pair. Note the prominent specularities. try of every bidirectional reflectance distribution function (BRDF) and is known as Helmholtz reciprocity [9, 15]. So while specular highlights generally move over the surface for changes of viewpoint under fixed lighting, in recipro- cal pairs they correspond to the projection of fixed surface points and essentially become features for determining cor- respondence. As originally formulated in [14, 21], stereo correspon- dence is established using a rank constraint that requires at least three reciprocal image pairs, and in turn this implies at least three camera locations and six images. (In fact, 36 images were used to great effect in [21].) Once correspon- dence is established, the same constraints can be used to estimate the normal field of the surface without the need for differentiation. In this paper, we re-examine the constraint arising from a single reciprocal pair of images and reveal that it defines a partial differential equation (PDE). We show that this PDE can be solved to provide an accurate reconstruction of sur- face shape. This new reconstruction technique, binocular Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE
Transcript
Page 1: Binocular Helmholtz Stereopsis

Binocular Helmholtz Stereopsis

Todd E. Zickler� Jeffrey Ho� David J. Kriegman� Jean Ponce� Peter N. Belhumeur�

[email protected] [email protected] [email protected] [email protected] [email protected]

�Electrical Engineering, Yale University, New Haven, CT 06511�Computer Science and Engineering, University of California at San Diego, La Jolla, CA 92093

�Computer Science, University of Illinois at Urbana-Champaign, Urbana, IL 61801�Computer Science, Columbia University, New York, NY 10027

Abstract

Helmholtz stereopsis has been introduced recently as a sur-face reconstruction technique that does not assume a modelof surface reflectance. In the reported formulation, corre-spondence was established using a rank constraint, neces-sitating at least three viewpoints and three pairs of images.Here, it is revealed that the fundamental Helmholtz stereop-sis constraint defines a nonlinear partial differential equa-tion, which can be solved using only two images. It is shownthat, unlike conventional stereo, binocular Helmholtz stere-opsis is able to establish correspondence (and thereby re-cover surface depth) for objects having an arbitrary andunknown BRDF and in textureless regions (i.e., regions ofconstant or slowly varying BRDF). An implementation andexperimental results validate the method for specular sur-faces with and without texture.

1 Introduction

Helmholtz stereopsis has recently been introduced as amulti-view technique for estimating surface shape withoutrequiring an assumed model of reflectance. Thus, unlikemost existing methods, it enables the dense reconstruc-tion of scenes that contain surfaces with unknown, spatiallyvarying, and arbitrary reflectance functions (BRDFs). InHelmholtz stereopsis, this is accomplished through the useof images that are collected in reciprocal pairs [21]. Thefirst image in a reciprocal pair is acquired under a singlepoint light source, and the second image is acquired after in-terchanging the camera and light source positions. Figure 1shows a reciprocal pair, and Fig. 2 illustrates the acquisi-tion geometry. Reciprocal pairs have the unique propertythat the relation between intensities at corresponding imagepoints depends only on surface shape and is independent ofreflectance. This property follows directly from the symme-

Figure 1. Two rectified images of a painted, plasticmannequin head acquired as a Helmholtz reciprocalpair. Note the prominent specularities.

try of every bidirectional reflectance distribution function(BRDF) and is known as Helmholtz reciprocity [9, 15]. Sowhile specular highlights generally move over the surfacefor changes of viewpoint under fixed lighting, in recipro-cal pairs they correspond to the projection of fixed surfacepoints and essentially become features for determining cor-respondence.

As originally formulated in [14, 21], stereo correspon-dence is established using a rank constraint that requires atleast three reciprocal image pairs, and in turn this impliesat least three camera locations and six images. (In fact, 36images were used to great effect in [21].) Once correspon-dence is established, the same constraints can be used toestimate the normal field of the surface without the need fordifferentiation.

In this paper, we re-examine the constraint arising froma single reciprocal pair of images and reveal that it defines apartial differential equation (PDE). We show that this PDEcan be solved to provide an accurate reconstruction of sur-face shape. This new reconstruction technique, binocular

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE

Page 2: Binocular Helmholtz Stereopsis

Helmholtz stereopsis, offers the following two significantadvantages over conventional stereopsis.

1. Binocular Helmholtz stereopsis is able to reconstructsurfaces with arbitrary and unknown BRDF’s (in-cluding very specular surfaces), whereas conventionaldense stereo correspondence is predicated on a con-stant brightness assumption (i.e., that the BRDF isLambertian).

2. Binocular Helmholtz stereopsis is able to establish cor-respondence in textureless regions, whereas conven-tional stereo can only “guess” correspondence in suchregions using a regularization or smoothing process.

The skeptical reader might turn to the reconstructions inFigures 5–7 of a specular mannequin shown in Fig. 1 andnotice the lack of texture on the forehead.

Additionally, by using only two images, binocularHelmholtz stereopsis is faster, simpler, and cheaper to im-plement than the multinocular Helmholtz stereo techniquereported in [21], and so it may be possible to apply this newbinocular technique within a much broader range of appli-cations.

Helmholtz stereopsis in general is related to a small set offairly recent reconstruction techniques (others are [13, 14])that use both changing viewpoint and illumination to re-construct surfaces with arbitrary and unknown BRDF. Luand Little [13] used the term photogeometric to describetheir technique, which seems like an appropriate term forthe entire class of methods. Helmholtz stereopsis differsfrom these methods, however, in that it both provides di-rect surface normal estimates and can handle a BRDF thatvaries over the surface. Conventional photometric stereocan also be applied to surfaces with a non-LambertianBRDF [10], and it is possible to reconstruct a surface fromonly two images [17]. However, the reflectance map (i.e.,BRDF) must be known a priori or be of a known parametricform [11, 19]. On the other hand, conventional stereo hasbeen augmented to handle specularities treating them as anoutlier process or using more than two views [3, 5].

In the next section, we show how the Helmholtz con-straint arising from a reciprocal pair leads to a differentialequation that, given some initial correspondence, can be in-tegrated across corresponding epipolar lines. (See also therecent work of Tu and Mendonca [20].) In Sec. 3, binoc-ular Helmholtz stereo is recast using a functional that doesnot require initial conditions; this functional is optimizedusing a multipass, dynamic programming algorithm. Ex-perimental results shown in Sections 2 and 3 demonstratethat accurate reconstruction is possible for non-Lambertiansurfaces. Conclusions and future directions are consideredin Sec. 4.

n̂ n̂

ol or ol or

l̂v vr̂vl̂ vr̂

p p

Figure 2. The binocular Helmholtz stereo setup.First an image is acquired with the scene illuminatedby a single point source as shown on the left. Then,a second image is acquired after the positions of thecamera and light source are exchanged as shown onthe right.

2 A PDE for Helmholtz Stereopsis

Consider the imaging geometry shown in the left half ofFig. 2 in which a scene is illuminated by an isotropic pointsource and observed by a perspective camera. Let �� and�� denote the positions of the camera and light source, re-spectively. We also denote by � and �� a point on the sur-face and its associated unit normal vector. The unit vectors��� �

�������

������ and ��� ��

������������ denote the di-

rections from � to the camera and light source, respectively.Given this system, the image irradiance at the projection of� is

�� � ������� ������ � ���

��� � ���(1)

where �� � ��� gives the cosine of the angle between the di-rection to the light source and the surface normal, �

�������

is the ���� fall-off from a unit-strength, isotropic point lightsource, and �� is the BRDF.

Now, consider the reciprocal case shown on the right ofFig. 2 in which the light source is positioned at ��, and thecamera observes � from ��. Because of Helmholtz reci-procity, we have that ������� ���� � ������� ����� This allowsus to eliminate the BRDF, and obtain the Helmholtz stereoconstraint first introduced in [14]:

���

������

��� � ���� ��

������

��� � ���

�� �� � � (2)

Note that Eq. 2 is a first order, nonlinear partial differen-tial equation in the point coordinates � and their derivativesas expressed through the normal ��. To solve this, we willfirst consider an imaging situation in which the PDE is sim-pler and for which the results are more transparent, and thenwe will impose the epipolar geometry.

Let the distances of the light source and camera to thescene be large with respect to the relief of the scene, and

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE

Page 3: Binocular Helmholtz Stereopsis

let the camera field of view be narrow. Under these condi-tions, the cameras can be modeled by scaled orthographicprojection, and the vectors ������ and ������ can be taken asconstant over the scene. As well, the denominators �������

and ��� � ��� can each be taken as constant over the scene.The ratio ������

������can be easily determined when calibrating

a Helmholtz stereo rig, and here we take this ratio to be �.Under these assumptions, Eq. 2 reduces to

������ � ������ � �� � �� (3)

where ��� and ��� are constants determined during calibra-tion.

We now impose the epipolar constraint to provide a so-lution to Eq. 3. Without loss of generality, establish a co-ordinate system for the left camera with a rotation matrixthat is the identity, so that ��� � ��� �����. Let the coordi-nates of points in the world be expressed in this system as��� �� �� where ��� �� are the image coordinates, and � is thedepth. We will consider the surface to be the graph of a �

function ���� ��� Furthermore, consider the pair of imagesto be rectified, so that the second camera’s orientation canbe expressed as

�� � ��������� �

�� ��� � ���

� � ����� � ����

�� � (4)

So, a point ��� �� �� will project to ��� �� in the left im-age and to ����� � ����� �� in the right image (i.e., thescan lines are epipolar lines). The disparity is then given by����� ������� ��� �.

Expressing the depth as ���� �� and noting that the un-normalized surface normal is � ��

��� ��������, we can write

the constraint as

����� ��� ������ ���������

��� ����� � � (5)

where �� � ���� � ������� ��, and this holds for all �.Rewriting this, we have

��

��� �

����� �� � ������ ���������������� ��

(6)

This can be numerically integrated as

���� �� �

� �

��

��

�� � � ����� ��� (7)

In other words, for each epipolar line �, this integral canbe independently evaluated to provide the depth across theepipolar line. Note that there is no search for correspon-dence over some disparity space, as correspondence is de-termined as a byproduct of integration. The only open issueis: “For each epipolar line �, what is the boundary condi-tion ����� ��?” There are two ways to look at this issue.

0 50 100 150 200 250 300 350−3

−2

−1

0

1

2

3

4

5

Figure 3. A family of reconstructions for one epipo-lar line of the specular cylinder shown in the bottomrow of Fig. 4. The family arises from different initialconditions ����� �� when integrating Eq. 7. The thick(red) curve is the member of this family with the cor-rect geometry, and is redrawn with a different scalingin the lower right of Fig. 4.

On the one hand, knowing ����� �� for some ���� �� oneach epipolar line amounts to having the means to deter-mine the depth or establish correspondence for one pointon each epipolar line. Alternatively, one can view Eq. 7as defining a one-parameter family of reconstructed curvesalong each epipolar line; elements of the family are indexedby different depth values at ���� ��. In Sec. 3, we will in-troduce a method for selecting a member of this family foreach epipolar line.

We have implemented this method in Matlab usingRunge-Kutta integration and empirically validated its effec-tiveness in the following experiment. We gathered recipro-cal pairs of images of three cylinders made of a Lamber-tian material, a rough non-Lambertian material [18], and aspecular plastic material. Images were acquired with a Ko-dak DCS 760 digital camera, and the scene was illuminatedwith a 150W halogen bulb. The camera system was geo-metrically calibrated, and the distance from the camera tothe object was about two meters which satisfies the approx-imation needed in deriving Eq. 3. Figure 4 shows for eachcylinder a pair of rectified images and a plot of the imageintensity across a pair of central epipolar lines. Note thatthese curves are characteristic of these three material types.

For the epipolar line, a family of reconstructions can beobtained for a discrete set of initial depths (disparities), andFig. 3 shows such a family for the specular cylinder. Since���� �� must be �, the integration cannot cross the occlud-ing contour where there is a depth (zeroth-order) discon-tinuity. Within this family lies the correct reconstruction,which could be selected from a single correspondence or

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE

Page 4: Binocular Helmholtz Stereopsis

some other means (e.g., smoothness, a shape prior, the dy-namic programming method introduced in Sec. 3, etc.). Thelast column of Fig. 4 shows the reconstructed depth acrossone epipolar line overlaid on a circular cross section. Inthis experiment, the initial conditions were chosen manu-ally. The RMS errors between the reconstructed curve andoverlaid circle as a percentage of the cylinder’s radius are0.11%, 1.7%, and 0.94% respectively for the Lambertian,generalized Lambertian, and specular cylinders. Note thatthe reconstructed curve for the Lambertian cylinder is indis-tinguishable from the ground truth circle whereas there is aslight deviation for the specular cylinder.

3 Surface Reconstruction

As discussed in the previous section, we can solve thebinocular Helmholtz PDE by integrating along correspond-ing epipolar lines. The recovery of the correct solution,however, requires a correct initial correspondence for start-ing the integration. This requirement may seem like a dif-ficult obstacle for the method; yet, we show in this sectionthat this can be easily overcome by applying a matchingconstraint common to traditional stereo algorithms.

Recall that traditional dense stereo algorithms must de-termine the correspondence for all points along correspond-ing epipolar lines in the left and right images. In contrast,binocular Helmholtz stereo needs only – in theory at least –to determine the correspondence of a single pair of points.This correspondence provides an initial condition which canthen be integrated to the left and right along the epipolar lineto determine all correspondences and establish depth for allpoints.

One could develop an algorithm that first identified themost prominent/salient feature (e.g., edges) for each pair ofepipolar lines and then assigned correspondence betweenthese features as way of creating anchor points for the in-tegration described above. Even for surfaces with arbi-trary BRDF, there will generically exist image intensity dis-continuities (edges) corresponding to discontinuities in thealbedo or BRDF across the surface, discontinuities in thesurface normal (creases), and discontinuities in depth at oc-clusion boundaries [4]. Consider more closely the two im-ages of the mannequin in Fig. 1. While the specular high-lights clearly indicate that this surface is far from Lamber-tian, there are also BRDF discontinuities arising from theeyes, eyebrows, and lips which can be used as features.

While coupling a feature-based stereo algorithm with theintegration method of Sec. 2 might work, it would rely onthe identification of feature points and would break downungracefully when the initial correspondences were incor-rectly assigned. Instead, we set up the problem of depthrecovery along epipolar lines as the minimization of a func-tional that includes a term for matching image features anda term for keeping the solution close to that dictated by the

integration. The functional is chosen to allow for efficient,global minimization using dynamic programming.

Pass 1: Along Epipolar Lines

As before, we assume that we have a reciprocal pair of or-thographic images taken of a � surface (a graph ���� ��),that the camera/source positions are far from the surface,and that the images are rectified. In Sec. 2, a differentialequation was derived for ��

��as a function of intensities at

corresponding points along an epipolar line, here we mod-ify the equation so that it is defined relative to a cyclopeancoordinate system as in [1].

Let ���� �� denote the following ratio of image measure-ments

���� �� � � � ���������� ������

������ � ������(8)

where � is the cyclopean coordinate; � is the half anglebetween viewing direction; �� � � ���� � � � ���� and�� � � � ��� � � � ���� are coordinates on the left andright epipolar lines, respectively. For corresponding points,we have that

��

��� ���� ��� (9)

The functional that we minimize has as its matching termthe squared magnitude of the difference between the gradi-ent in the left and right images. This matching term is cou-pled with a term measuring the squared distance between���� �� and the partial derivative of the hypothesized solu-tion � with respect to �. The functional is given as

�������� �

����

��� ��� ���

����������

� � (10)

where �� � �������� ���, �� � �������� ���, and � is aweighting term. For each epipolar line, we use dynamicprogramming to find a discrete approximation to ���� thatminimizes ��������. For � discrete values of �, we con-sider � possible depth values, and the computational costof finding the global minimum of �������� is ������; see[2] for details.

Pass 2: Across Epipolar Lines

If each epipolar line had an image feature such as an albedoedge, the gradient matching term would lock onto the edgeand effectively provide the initial correspondence neededfor the integration. We are not guaranteed, however, thateach epipolar line will have an image feature that will allowthe minimization to find the correct initial correspondenceand, thus, the correct solution. Reconsider the images of themannequin in Fig. 1. While there are strong image featuresfor epipolar lines crossing the eyebrows, eyes and mouth re-gions, the forehead is smooth and textureless, whereas thenose has significant relief and little texture. However, we

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE

Page 5: Binocular Helmholtz Stereopsis

0 100 200 300 400 500 6000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

−0.2 0 0.2 0.4 0.6 0.8 1

0

0.2

0.4

0.6

0.8

1

0 100 200 300 400 500 6000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

−0.2 0 0.2 0.4 0.6 0.8 1

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

0 100 200 300 400 500 600 700 8000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

−0.2 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8

0

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

Figure 4. Reconstruction of three real cylinders of three material types: The cylinder in Row 1 is approximatelyLambertian, the rough cylinder in Row 2 has Oren-Nayar generalized Lambertian reflectance [18], and the plasticcylinder in Row 3 is highly specular. The first two columns show a rectified pair of images of the cylinder. The thirdcolumn shows a plot of the image intensities across one epipolar line in the left (blue, solid) and right (red, dashed)images. The fourth column shows the reconstructed shape (thick, blue) across the epipolar line superimposed on acircular cross section.

expect that in general other epipolar lines will have imagefeatures, and we can use these to determine the best solu-tions for epipolar lines that do not. To operationalize this,we define and minimize a second stage functional acrossepipolar lines (in a manner different than that used in [16]).

The idea is to compute as the output from Pass 1 a familyof � solutions minimizing Eq. 10 such that the endpointsof the solutions vary over a range of possible � values. Ifthe family of solutions is big enough (i.e., if our samplingof the range of � values is fine enough), then the correctsolution should be well represented by one member fromthe family of solutions. Note that this should hold whetheror not the epipolar line has an image feature. To choose thecorrect solution for each epipolar line, we simply choosethe collection of solutions (one for each line) that producesthe most coherent surface according to a simple smoothnesscriteria imposed across epipolar lines.

More precisely, let � � � denote the end of the epipolarline. Let ���� �� � � denote ���� �� at the endpoint � �

� for scanline �. For each � and for each ending point �in the range of possible � values, we compute a solution

����� ������ ����� � ���� ��������������������

������� ���� (11)

In other words, ����� ������ �� � �� is the solution alongepipolar line � that minimizes Eq. 10 subject to the con-straint that ���� � �. Thus for each �, the family of so-lutions is indexed by the value of the ending point �. Notethat this family differs from the one arising in Sec. 2 andshown in Fig. 3. And, within this family there should be a� and a corresponding solution ����� ������ �� � �� thatis close to the correct solution. We denote the family of so-lutions over all epipolar lines by �� � ������ ������� �� ���� ������.

Nevertheless, we still have the problem of determiningthe correct solution from �� for each epipolar line �. Con-

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE

Page 6: Binocular Helmholtz Stereopsis

Figure 5. Two views of a reconstructed plastic mannequin created using binocular Helmholtz stereopsis.

sider the following functional defined over the entire image

������ ��� �

� � ���

��

��

� �� (12)

We can find the correct choice from �� for each epipolar line� by minimizing the above functional. We do not considerall possible solutions ���� ��, rather we limit the search ofsolutions to the family of solutions �� given by Pass 1. Thus,the optimization in Eq. 12 is effectively being done over theendpoints � for each �. We take as our solution to Pass 2,

����� �� � ���� ��

������ ���� (13)

As in Pass 1, we use dynamic programming to finda discrete approximation to the ����� �� that minimizes������ ���. The computational cost of this dynamic pro-gramming step is ������ where � is the number of end-points (depth values), and � is the number of epipolar lines.

Note that this formulation has an inherent asymmetry, asthe second pass considers a range of ending points and nota range of starting points. We correct this by re-running thistwo stage process in reverse. Specifically, we run Pass 1 andPass 2 across the data to find a collection of optimal endingpoints �� � ����� �� for each epipolar line. We then re-runPass 1 in reverse (i.e., from right to left), fixing the endpointsuch that ���� �� � �� for each �. At this stage, for each �we now have a family of solutions indexed by the value ofthe beginning point ��. The overall solution is then chosenby re-running Pass 2 to select the optimal starting points.

We should point out that this algorithm has only one realparameter: the weighting � of the image gradient term usedin optimization of Eq. 10. The optimization in Eq. 12 is pa-rameter free. This optimization does not smooth the solu-tion along the epipolar lines, rather it chooses the solutionswhich together form the surface that is smoothest across theepipolar lines.

Here, we present results on a reciprocal pair of images ofa mannequin head shown in Fig. 1. In Fig. 5, we display two

views of a mesh of the surface reconstructed using our twopass dynamic programming method. For this reconstruc-tion, the brightness in the left and right images was normal-ized by the maximum brightness in both images, and thevalue of � � ���. In Fig. 6, we display a depth map imagein which light corresponds to near and dark to far. Finally,in Fig. 7, we display a single view of the mesh with theleft image texture mapped onto it. Notice that the methodis unhampered by the specularities and is able to both “lockonto” the features such the eyes, eyebrows, and lips, but alsoprovide good reconstructions in textureless regions such asthe forehead.

4 Discussion

This paper introduces binocular Helmholtz stereopsis andshows that the constraint arising from reciprocal images isa partial differential equation which can be readily solved.The implementation validates that the method can be ap-plied to objects with arbitrary and unknown BRDF, and un-like conventional stereo it is able to accurately reconstructshape in textureless regions.

There are still many ways to improve and enhance binoc-ular Helmholtz stereopsis. First, it is straightforward toextend the implementation to perspective projection andnearby light sources, though the PDE becomes more com-plex. Second, it is assumed throughout that the depth func-tion ���� �� is �, yet it should be possible to permit depthdiscontinuities. As described in [14, 21], the equivalent tohalf-occluded regions corresponding to depth discontinu-ities in conventional stereo are shadowed regions in recip-rocal image pairs. Hence, it should be possible to augmentEq. 10 to introduce depth discontinuities as in [1, 6, 8], us-ing shadows as a cue for half occlusion.

Finally, though an advance brought forth in this paper isto reduce Helmholtz stereopsis from requiring at least threecamera positions [14, 21] to binocular imaging, it would beworthwhile to re-examine multinocular Helmholtz stereop-sis by directly considering the set of differential equationsarising from each reciprocal pair. It may then be possible to

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE

Page 7: Binocular Helmholtz Stereopsis

Figure 6. A depth map from the cyclopean viewpointof the mannequin face in which light corresponds tonear and dark to far.

Figure 7. A texture mapped reconstruction of themannequin face, rendered with a Phong reflectancemodel.

eliminate the smoothness constraint between epipolar linesused in the second pass of Sec. 3. Perhaps, multinocularHelmholtz reconstruction can be formulated to exploit ad-vantages found through photoconsistency and space carv-ing [12] or level set methods [7].

Acknowledgements

P.N. Belhumeur and T. Zickler were supported by theNational Science Foundation under grants PECASE IIS-9703134, ITR IIS-00-85864, EIA-02-24431, IIS-03-08185,and KDI-99-80058. D.J. Kriegman and J. Ho were sup-ported by the NSF under grants EIA-00-04056, CCR-00-86094 and IIS-03-08185. J. Ponce was supported by theUIUC Campus Research Board and the NSF under grantIIS-03-08087.

References

[1] P. Belhumeur. A Bayesian approach to binocular stereopsis.Int. Journal of Computer Vision, 19, 1996.

[2] R. Bellman. Dynamic Programming. Princeton UniversityPress, 1957.

[3] D. Bhat and S. Nayar. Stereo and specular reflection. Int. J.Computer Vision, 26(2):91–106, 1998.

[4] T. Binford. Generic surface interpretation: Observabilitymodel. In Proc. of the �

�� International Symposium onRobotics Research, Santa Cruz, CA, August 1987.

[5] A. Blake and G. Brelstaff. Specular stereo. In Proc. Int. JointConf. on Art. Intell., pages 973–976, 1985.

[6] I. J. Cox, S. Hingorani, B. M. Maggs, and S. B. Rao. Stereowithout disparity gradient smoothing: a Bayesian sensor fu-sion solution. In D. Hogg and R. Boyle, editors, British Ma-chine Vision Conference, pages 337–346. Springer-Verlag,1992.

[7] O. Faugeras and R. Keriven. Complete dense stereovisionusing level set methods. In Proc. European Conf. on Com-puter Vision, 1998.

[8] D. Geiger, B. Ladendorf, and A. Yuille. Occlusions in binoc-ular stereo. In Proc. European Conf. on Computer Vision,Santa Margherita Ligure, Italy, 1992.

[9] H. v. Helmholtz. Treatise on Physiological Optics, volume 1.Dover (New York), 1925.

[10] B. Horn. Robot Vision. MIT Press, Cambridge, Mass., 1986.[11] K. Ikeuchi and B. Horn. Numerical shape from shading and

occluding boundaries. Artificial Intelligence, 17:141–184,1981.

[12] K. Kutulakos and S. Seitz. A theory of shape by space carv-ing. Int. J. Computer Vision, 38(3):197–216, July 2000.

[13] J. Lu and J. Little. Reflectance and shape from images usinga collinear light source. Int. Journal of Computer Vision,32(3):1–28, 1999.

[14] S. Magda, T. Zickler, D. Kriegman, and P. Belhumeur.Beyond Lambert: Reconstructing surfaces with arbitraryBRDFs. In Proc. Int. Conf. on Computer Vision, pages II:391–398, 2001.

[15] F. Nicodemus, J. Richmond, J. Hsia, I. Ginsberg, andT. Limperis. Geometric considerations and nomenclature forreflectance. Monograph 160, National Bureau of Standards(US), 1977.

[16] Y. Ohta and T. Kanade. Stereo by intra- and inter-scanlinesearch. IEEE Transactions on Pattern Analysis and MachineIntelligence, 7(2):139–154, 1985.

[17] R. Onn and A. Bruckstein. Integrability disambiguates sur-face recovery in two-image photometric stereo. Int. J. Com-puter Vision, 5(1):105–113, 1990.

[18] M. Oren and S. Nayar. Generalization of the Lambertianmodel and implications for machine vision. Int. J. ComputerVision, 14:227–251, 1996.

[19] H. Tagare and R. deFigueiredo. A theory of photomet-ric stereo for a class of diffuse non-lambertian surfaces.IEEE Trans. Pattern Anal. Mach. Intelligence, 13(2):133–152, February 1991.

[20] P. Tu and P. Medonca. Surface reconstruction via Helmholtzreciprocity with a single image pair. In Proc. IEEE Conf.Computer Vision and Pattern Recognition, June 2003.

[21] T. Zickler, P. Belhumeur, and D. Kriegman. Helmholtz stere-opsis: Exploiting reciprocity for surface reconstruction. InProc. European Conf. on Computer Vision, pages III: 869–884, May 2002.

Proceedings of the Ninth IEEE International Conference on Computer Vision (ICCV 2003) 2-Volume Set 0-7695-1950-4/03 $17.00 © 2003 IEEE


Recommended