+ All Categories
Home > Documents > Semantically-Aware Aerial Reconstruction from Multi-Modal...

Semantically-Aware Aerial Reconstruction from Multi-Modal...

Date post: 23-May-2020
Category:
Upload: others
View: 3 times
Download: 0 times
Share this document with a friend
9
Semantically-Aware Aerial Reconstruction from Multi-Modal Data Randi Cabezas Julian Straub John W. Fisher III Massachusetts Institute of Technology {rcabezas,straub,fisher}@csail.mit.edu Abstract We consider a methodology for integrating multiple sen- sors along with semantic information to enhance scene representations. We propose a probabilistic generative model for inferring semantically-informed aerial recon- structions from multi-modal data within a consistent math- ematical framework. The approach, called Semantically- Aware Aerial Reconstruction (SAAR), not only exploits in- ferred scene geometry, appearance, and semantic obser- vations to obtain a meaningful categorization of the data, but also extends previously proposed methods by impos- ing structure on the prior over geometry, appearance, and semantic labels. This leads to more accurate reconstruc- tions and the ability to fill in missing contextual labels via joint sensor and semantic information. We introduce a new multi-modal synthetic dataset in order to provide quanti- tative performance analysis. Additionally, we apply the model to real-world data and exploit OpenStreetMap as a source of semantic observations. We show quantitative im- provements in reconstruction accuracy of large-scale urban scenes from the combination of LiDAR, aerial photogra- phy, and semantic data. Furthermore, we demonstrate the model’s ability to fill in for missing sensed data, leading to more interpretable reconstructions. 1. Introduction Humans integrate various sensory, semantic, and contex- tual cues to construct internal representations of the world for robustly reasoning within their environment. This abil- ity is little diminished in the face of sparse and noisy obser- vations. Furthermore, humans easily extrapolate the struc- ture of the surrounding large-scale environment from their local surroundings using cues and prior experiences of sim- ilar scenes. Such inferential feats are supported by semantic understanding and categorization of scene elements. Moti- vated by such abilities we develop an approach for scene reconstruction that integrates sensor data along with geo- referenced semantic labels. With some exceptions, existing approaches focus solely on obtaining a semantic labeling Figure 1: Top: Lubbock scene inferred geometry and labels. Bottom: Aerial image, LiDAR and OSM observations. from sensor data. We construct a probabilistic generative model that allows multi-modal data fusion, integrates se- mantic observations, and captures the notion that different scene components exhibit different properties via a struc- tured prior over the different modalities. The approach, Semantically-Aware Aerial Reconstruction (SAAR), infers semantically-meaningful categories of data which are used in the reconstruction as category-specific priors, resulting in improved reconstruction accuracy. SAAR builds on the work of Cabezas et al.[4], by adding semantic observations from OpenStreetMap (OSM) [12] to the sensor data already used there: aerial oblique imagery, and Light Detection and Ranging (LiDAR). Each of these sources provides complementary features for rea- soning about urban scenes. We further expand on [4] by in- troducing a novel structured prior, to regularize reconstruc- tions. In contrast to prior semantic reconstruction work, our goal is to (1) exploit the learned categories to improve the reconstruction and (2) fill in missing sensor data. Previous work in the area of semantic labeling can be categorized based on the domain of the labels: image-space or 3D-space. In the first category, Sudderth et al.[31] and Wang et al.[35] focus on the extraction of pixel-wise labels
Transcript
Page 1: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

Semantically-Aware Aerial Reconstruction from Multi-Modal Data

Randi Cabezas Julian Straub John W. Fisher IIIMassachusetts Institute of Technology

{rcabezas,straub,fisher}@csail.mit.edu

Abstract

We consider a methodology for integrating multiple sen-sors along with semantic information to enhance scenerepresentations. We propose a probabilistic generativemodel for inferring semantically-informed aerial recon-structions from multi-modal data within a consistent math-ematical framework. The approach, called Semantically-Aware Aerial Reconstruction (SAAR), not only exploits in-ferred scene geometry, appearance, and semantic obser-vations to obtain a meaningful categorization of the data,but also extends previously proposed methods by impos-ing structure on the prior over geometry, appearance, andsemantic labels. This leads to more accurate reconstruc-tions and the ability to fill in missing contextual labels viajoint sensor and semantic information. We introduce a newmulti-modal synthetic dataset in order to provide quanti-tative performance analysis. Additionally, we apply themodel to real-world data and exploit OpenStreetMap as asource of semantic observations. We show quantitative im-provements in reconstruction accuracy of large-scale urbanscenes from the combination of LiDAR, aerial photogra-phy, and semantic data. Furthermore, we demonstrate themodel’s ability to fill in for missing sensed data, leading tomore interpretable reconstructions.

1. IntroductionHumans integrate various sensory, semantic, and contex-

tual cues to construct internal representations of the worldfor robustly reasoning within their environment. This abil-ity is little diminished in the face of sparse and noisy obser-vations. Furthermore, humans easily extrapolate the struc-ture of the surrounding large-scale environment from theirlocal surroundings using cues and prior experiences of sim-ilar scenes. Such inferential feats are supported by semanticunderstanding and categorization of scene elements. Moti-vated by such abilities we develop an approach for scenereconstruction that integrates sensor data along with geo-referenced semantic labels. With some exceptions, existingapproaches focus solely on obtaining a semantic labeling

Figure 1: Top: Lubbock scene inferred geometry and labels.Bottom: Aerial image, LiDAR and OSM observations.

from sensor data. We construct a probabilistic generativemodel that allows multi-modal data fusion, integrates se-mantic observations, and captures the notion that differentscene components exhibit different properties via a struc-tured prior over the different modalities. The approach,Semantically-Aware Aerial Reconstruction (SAAR), inferssemantically-meaningful categories of data which are usedin the reconstruction as category-specific priors, resulting inimproved reconstruction accuracy.

SAAR builds on the work of Cabezas et al. [4], byadding semantic observations from OpenStreetMap (OSM)[12] to the sensor data already used there: aerial obliqueimagery, and Light Detection and Ranging (LiDAR). Eachof these sources provides complementary features for rea-soning about urban scenes. We further expand on [4] by in-troducing a novel structured prior, to regularize reconstruc-tions. In contrast to prior semantic reconstruction work, ourgoal is to (1) exploit the learned categories to improve thereconstruction and (2) fill in missing sensor data.

Previous work in the area of semantic labeling can becategorized based on the domain of the labels: image-spaceor 3D-space. In the first category, Sudderth et al. [31] andWang et al. [35] focus on the extraction of pixel-wise labels

Page 2: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

from single images with no additional information. Alongsimilar lines, Cao et al. [5] and Choi et al. [6] expand onprior work by introducing spatial connectivity between thelabels. They show that spatial connectivity greatly improvesperformance. Similarly, the problem of matching noisy texttags to images and, if possible, identifying image regionsthat give rise to the textual description has been extensivelystudied [8, 19, 20, 22, 24, 25, 38]. These approaches typ-ically learn multi-modal (text and appearance) representa-tions and attempt to predict one modality when the other ismissing. All of the aforementioned works are formulated inthe image-domain and differ from the proposed approach inboth formulation and goal.

Prior work in 3D scene labeling can be categorized bychoice of primitive used to produce the labeling: point,voxel, mesh, or object box (while not strictly a primi-tive, it is included for completeness). Point-based se-mantic labeling has received the most attention in recentyears [1, 3, 10, 14, 18, 37]. The goal is to label 3D points, ei-ther using human annotations and propagating them throughthe point cloud, or by following procedures similar toimage-based methods and projecting the results into 3D us-ing standard multi-view techniques. Voxel-based methods,e.g., [13, 16, 17, 28, 34], formulate the labeling problem asan energy minimization problem in either 2D (followed byprojection to 3D), or directly in 3D. Most of these methodsexploit spatial connectivity using Markov Random Fieldsor Conditional Random Fields (CRFs). Like some voxel-based methods, mesh-based approaches [2, 33] rely on en-ergy minimization in a CRF; however, unlike voxel-basedapproaches, they tend to have a richer set of discrimina-tive features such as texture, curvature and various meshproperties. Approaches that label scene objects (typicallyin the form of a bounding box) provide a higher degree ofabstraction than primitive-based methods. Ren et al. [27]and Lin et al. [21] rely on a set of low-level features to findand classify scene regions.

The proposed work falls in the mesh-based category.Like other methods in this category, it relies on high-levelfeatures, including primitive appearance, geometry via lo-cation and orientation, and, if available, semantic observa-tions, such as OSM or Geiger mode LiDAR. This data isused in a probabilistic model to learn scene categories andcategory-specific structured priors that can be used to regu-larize scene elements during reconstruction.

The contributions of this work are: (1) a novel proba-bilistic model that couples semantic labels and scene recon-struction, (2) mathematically consistent methods for obtain-ing labels and reconstructions from multi-modal data in 3D,(3) demonstration of the utility of semantic observationsand structured prior distributions to improve the accuracyof scene reconstruction, and (4) the introduction of a newphoto-realistic multi-modal synthetic urban city dataset.

ϕG ϕS π ϕA

θ Z

V G S A K T

L O I P

NP

NC

NINL NO

NV

NSNS

Figure 2: Graphical representation of the SAAR model.The parts taken over from [4] are depicted in gray whereasthe proposed structured prior is shown in black.

2. The Probabilistic SAAR Model

The proposed SAAR model couples a latent structuredprior model with a probabilistic semantic 3D world repre-sentation. The 3D representation is typically based on acollection of independent primitives (e.g., points, voxels, ortriangles as in our case) with a series of attributes that de-scribe the various scene aspects (e.g., geometry and appear-ance). SAAR draws from the model proposed in [4] where,in addition to the latent geometry and appearance per primi-tive (i.e., triangle), we also model a semantic label. Further-more, we introduce a structured prior via a mixture modelover the latent semantic labels, appearance, and geometryto replace the uninformative uniform priors used in [4]. Wewill show that posterior inference under such a prior cap-tures meaningful scene-specific global structure, which canbe leveraged to regularize the 3D reconstructions. These ex-tensions lead to powerful regularizers for 3D reconstructionwithout the need of carefully hand-crafted scene priors.

From the generative viewpoint, the SAAR model de-scribes how the combination of latent geometry G, appear-ance A, and semantic label S give rise to LiDAR measure-ments L, observed OSM labels O, and camera images I atobserved GPS positions P. The geometry is represented viavertex locations V and connectivity matrix θ. The imagesare assumed to be generated from a set of fly-by cameraswith poses T (extrinsic) and calibration K (intrinsic). Forconvenience we let W , {G,A,S,Z}, where Z is theprimitive’s categorical assignment. The probabilistic graph-ical representation of this model is visualized in Fig. 2 (hy-perparameters are omitted for clarity). Model parametersare summarized in Tab. 1. The joint distribution for theprobabilistic SAAR model is:

p(L, I,P,O,T,K,V,W, ϕ, π; θ) = p(W, ϕ, π|V; θ)

×NV∏v=1

p(Vv)

NL∏l=1

p(Ll|G)

NO∏o=1

p(Oo|S,G)

×NC∏c=1

p(T c)p(Kc)

NcI∏n=1

p(Icn|G,A,Kc, T c)p(P cn|T c) ,

(1)

Page 3: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

Variables DescriptionNP , NV , NL, NO, NC , {N c

I }NCc=1 Number of primitives, vertices, LiDAR and OSM points; cameras and images.

NS , ND, NA, NB Number of semantic and OSM categories; appearance and image pixels.Ll ∈ R3 LiDAR observation.

Oo = (Po ∈ R3, Co ∈ {1, ..., Nd}) OSM observation (location, categories).Vm ∈ R3, Gm ∈ N1×3, θ ∈ RNp×3 Vertex location, Geometric primitive (triangle), Connectivity matrix.

Am = {ap |ap ∈ R3}N2A

p=1 Primitive appearance (texture) and corresponding RGB pixel.Sm = {Co|∀o assigned to Gm} OSM category distribution (all observations assigned to primitive m).

Zm ∈ {1, ..., Ns} Primitive category.T c ∈ SE(3), Kc ∈ (0, 180] Extrinsic trajectory (position and orientation) and Intrinsic parameter (FOV).Icn = {pj |pj ∈ R3}NBj=1 nth image taken with camera c modeled as a collection of RGB pixels.

P cn ∈ R3 nth GPS observation of camera c.π, ϕ = {ϕG, ϕA, ϕS} Cluster proportions; geometry, appearance and semantic parameters (see text).

Table 1: List of variables used in the model.

where the structured prior over primitive geometry, appear-ance, and semantic label factors as:

p(W, ϕ, π|V; θ) = p(π)

NS∏k=1

p(ϕGk )p(ϕ

Ak )p(ϕ

Sk )

NP∏m=1

[p(Zm|π)

× p(Gm|ϕG, Zm,V; θ)p(Sm|ϕS , Zm)p(Am|ϕA, Zm)]. (2)

We now describe the terms in Eq. (1). The image like-lihood, p(Icn|G,A,Kc, T c) =

∏NBk=1 N (ik; am(k), r

2m(k)),

is modeled as a Gaussian distribution with mean corre-sponding to the latent appearance pixel of the associatedprimitive and the variance corresponding to the inverse ofthe dot product between the camera’s viewing direction andthe surface normal of the visible primitive. The LiDARlikelihood is given by p(Ll|G) = N (d2(Ll, Gm(l)); 0, σ2),where d(Ll, Gm) is the distance between the LiDAR pointand primitive. The GPS observations follow a Gaussian dis-tribution centered at the camera trajectory, T c. All priors,with the exception of the structured prior, are uniform. Forfurther details see [4].

Semantic observations O are modeled in SAAR as a col-lection of independent 3D points, where each point has alocation and semantic label, Oo , (Po, Co). The locationand label are assumed to be independent, thus the likelihoodis p(Oo|S,G) = p(Po|G)p(Co|S). The location model isthe same as that of the LiDAR measurements; i.e., the like-lihood depends on the distance between the point and thegenerating primitive. The class likelihood is modeled as aCategorical distribution (Cat): p(Co|S) = Cat(Co;Sm(o)).

Equation (2) shows the mixture model structure of theprior on geometry, appearance, and semantic label, shownin black in Fig. 2. Note that each primitive m is assigned toa single mixture component via Zm. Each mixture compo-nent defines a distribution over the product space of geom-etry, appearance and semantic labels. Following the stan-dard approach in Bayesian mixture modeling, we assume a

Dirichlet distribution (Dir) prior with hyperparameter α onthe mixture weights π. The class labels Zm are distributedaccording to a categorical distribution parametrized by π:p(Zm|π) = Cat(Zm;π).

Conditioned on the category assignment via label Zm,the geometry, appearance, and semantic label of primitivem are modeled as generated independently from the associ-ated component distribution in the respective space. For thisprocess, we adopt a pixel-centric perspective inside each tri-angle. Specifically, we model the 3D location, surface nor-mal, RGB color and semantic label of each primitive’s ap-pearance pixels as independently and identically distributedaccording to the corresponding mixture component distri-bution. We note that this modeling is performed using thelatent primitive attributes. Under this model, we can col-lect the sufficient statistics over the different modalities anduse them for both likelihood evaluations and posterior in-ference. In the following, we introduce the distributions forappearance, semantic labels and geometry.

Appearance: For each mixture component, we modelthe appearance as a three-dimensional Gaussian in theRGB color space. Hence, given appearance parameterϕAk , (µAk ,Σ

Ak ) we have

p(Am|ϕAZm) =∏NAm

i=1N (Am,i; µ

AZm ,Σ

AZm) , (3)

where Am,i is the RGB color of pixel i in primitive m andthe product is over all pixels in the primitive. The Gaussianparameters of ϕA are distributed according to the NormalInverse Wishart (NIW) conjugate prior distribution [11].

Semantic labels: Labels Sm assigned to primitive mare modeled as following a categorical distribution with aDirichlet prior:

p(Sm|ϕSZm) =∏NAm

i=1Cat(Sm,i; ϕSZm) . (4)

Page 4: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

S2

TµS2

N (0,Σ)

Figure 3: Left: Tangent space Gaussian (TG) around meanµ with covariance Σ. Right: Manhattan Frame (MF).

Geometry: The geometry of primitive m is modeledvia the primitive’s appearance pixel locations in 3D, GXm,i,as well as its orientations, GQm,i, for pixel i. The pixel lo-cations GXm,i are assumed to be Gaussian distributed in 3Dwith NIW distribution priors on the Gaussian parameters{µGk ,ΣGk }, along with likelihood

p(GXm|ϕGZm) =∏NAm

i=1N (GXm,i; µ

GZm ,Σ

GZm) . (5)

As we will see, the structured prior uses the location infor-mation as a weak coupling between neighboring triangles.The orientation GQm,i of the primitive’s pixel i is describedvia its surface normal represented as a unit-length directionvector in 3D. The space of such vectors is the unit spherein 3D, S2. Accordingly, we model primitive orientations byplacing zero-mean Gaussian distributions in tangent spacesto S2 as explored in [29, 30]. We denote this distribution, vi-sualized in Fig. 3, the tangent space Gaussian (TG). Underthe TG model, surface normals GQm ∈ S2 associated withcluster Z , Zm have the following distribution

p(GQm|µQZ ,ΣQZ ) =

∏NAm

i=1N (LogµQZ

(GQm,i); 0,ΣQZ ) (6)

where the Riemannian logarithm map LogµQZ(GQm,i) maps

GQm,i into the tangent space TµQZS around the mean of the

TG µQZ ∈ S2. The TG model uses an Inverse Wishartprior [11] in the tangent plane for the covariance and a uni-form prior on the sphere for the mean. We explore twodifferent models: (1) an unconstrained model [29] with asingle TG per cluster and (2) the Manhattan Frame (MF)model [30]. The MF captures the block structure of man-made environments in the space of surface normals throughsix TG clusters that are constrained to orthogonal and op-posing locations on the sphere (Fig. 3).

3. InferenceIn this section we discuss sampling-based inference for

the SAAR model. We begin by outlining inference for thestructured prior portion of the model, followed by an infer-ence scheme for the full SAAR model.

Algorithm 1 Structured Prior Inference

1: Initialize ϕG, ϕA , ϕS and π from priors.2: for i ∈ {1, . . . , Niter} do3: Sample Z according to Eq. (7).4: Sample ϕG, ϕA ϕS and π according to Eq. (8).5: end for

3.1. Structured Prior Inference

As stated in the previous section, the structured prior forgeometry, appearance, and semantic labels is equivalent to amixture model over the aforementioned modes. This moti-vates the use of a Gibbs sampler that iterates between sam-pling labels Zm for each primitive and sampling mixturecomponent parameters for posterior inference. We presentthe necessary posterior distributions in the following; for adetailed derivation see Sup. Mat.

Label posterior: The posterior distribution for a labelZm of primitive m given the mixture parameters is

p(Zm = k|Z\m,G,A,S, π, ϕG, ϕA, ϕS)

∝ p(Am, Gm, Sm, ϕG, ϕA, ϕS |Zm)p(Zm = k|π)

= p(Gm, ϕG|Zm)p(Am, ϕ

A|Zm)p(Sm, ϕS |Zm)πk

∝ p(Gm|ϕGk )p(Am|ϕA

k )p(Sm|ϕSk )πk .

(7)

For each primitivem we can evaluate Eq. (7) under all clus-ters k using the likelihoods described in the previous sectionto obtain a discrete probability distribution. After normal-ization, we sample the indicator Zm.

Parameter posteriors: The posterior distributions overthe parameters of the mixture model factors into the differ-ent modes as

p(π, ϕG, ϕA, ϕS |Z,G,A,S)

∝ p(π|Z)p(ϕA|A,Z)p(ϕG|G,Z)p(ϕS |S,Z)

∝ p(π|Z)NS∏k=1

p(ϕAk |AIk )p(ϕ

Gk |GIk )p(ϕ

Sk |SIk )

(8)

where we use the indicator set Ik = {m : zm = k} tocollect all primitives that are assigned to cluster k. Due toconjugacy of the priors on π, ϕG, ϕA, and ϕS , the posteriorparameter distributions take the same form as the prior dis-tributions. This allows efficient sampling of posterior pa-rameters after updating the sufficient statistics [11]. Pos-terior sampling for the TG and MF distribution in ϕG iscarried out as described in [29] and [30] respectively. TheGibbs sampler for the structured prior is outlined in Alg. 1.

3.2. SAAR Model InferenceInference on the structured prior portion of the model is

coupled with the inference scheme proposed in [4] to sam-ple from the posterior of the SAAR model. Conceptually,the SAAR algorithm interleaves the geometry, appearance,

Page 5: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

Algorithm 2 Full SAAR Model Inference

1: Initialize world primitives and camera pose2: Sample assignment of LiDAR and OSM data.3: Initialize ϕG, ϕA, ϕS , π and Z4: for i ∈ {1, . . . , Nup} do5: Estimate appearance A.6: Optimize over camera pose T and K.7: Run Structured Prior inference procedure (Alg. 1).8: Estimate appearance A.9: Optimize over world primitive geometry, V.

10: Sample assignment of LiDAR and OSM data.11: end for

and camera pose sampling with posterior sampling of thestructured prior as outlined in Alg. 2. Specifically, the up-dates to the appearance and vertex locations now take thestructured priors into account. The appearance updates havethe same form as in [4]:

p(A|I,G,K,T,Z, ϕA) (9)

∝NC∏c=1

NcI∏n=1

NB∏k=1

p(In,ck |G,A,K

c, T c)

NP∏m=1

NA∏i=1

p(Am,i|ϕA, Zm) ,

where we now utilize the inferred appearance parameters ofthe assigned cluster. Since the form of the equation does notchange, we can still solve it in closed-form. Similarly, thevertex terms are augmented with the semantic label

p(V,G|I,L,O,Z,S, ϕG; θ) ∝No∏o=1

p(Oo|G,S)NL∏l=1

p(Ll|G)

×NC∏c=1

NcI∏n=1

p(Icn|G,A,Kc, T c)

NP∏m=1

p(Gm|ϕG,V;Z; θ) . (10)

We optimize over Eq. (10) using a downhill simplex opti-mization method [23] to update the latent vertex locationsto their maximum a posteriori configuration as in [4].

4. ResultsTo facilitate quantitative experiments, we created a

multi-modal photo-realistic urban city synthetic dataset,Synthetic City (SynthCity). In the following, we briefly de-scribe SynthCity and present several experiments to validatethe propose model. Experiments include an evaluation ofmodel performance in terms of reconstruction accuracy, andan ablation study to identify the best modalities for scenecategorization. Furthermore, we demonstrate the utility ofthe model by showing its ability to operate in the presenceof noisy data and to estimate missing sensor data. Finally,we present results on real-world scenes where we qualita-tively show the method’s improvements over the baselineof [4]. Throughout this section we’ll use the shorthand no-tation TG-N (or MF-N) to refer to a TG (or MF) orientationmodel with N semantic categories, i.e., NS = N .

Figure 4: Sample view of SynthCity dataset (City3) alongwith ground-truth mesh colored by semantic categories.

4.1. SynthCity Dataset

The lack of ground-truthed aerial datasets motivatesthe creation of the SynthCity dataset. We used Esri’sCityEngine [9] to create five randomly-generated realisticcities (ToyCity, ToyCity2, City1-1, City3, and City4), eachwith eight different types of scene elements: open space,streets, sidewalks, parking lots, residential buildings, officebuildings, high-rise buildings, and vegetation. Custom buildrules allowed pseudo-random variation of geometry and ap-pearance, resulting in a collection of elements that are notonly photo-realistic, but also closely match the layout ofreal-world cities (Fig. 4). LiDAR measurements were ob-tained by simulating real LiDAR collections [15]. Pleaserefer to the Sup. Mat. for further details.

4.2. Improved Reconstructions - SynthCity

We exploit the resulting categorization and priors to im-prove reconstruction accuracy via label-dependent updatesto geometry and appearance. The goal is to compare theeffect of the structured prior and semantic information onreconstruction accuracy. Fig. 5 compares various config-urations of SAAR (using all available data: appearance,location, orientation and semantic observations) with thenon-structured prior model [4], and with the LiDAR-onlymethod of Zhou et al. [39] on the SynthCity dataset. Thedistance metric used is the mean error between the esti-mated and true mesh as computed by Metro [7]. From thefigure we can see that in all cases SAAR produces betterreconstructions than both baseline approaches. We empha-size that the only difference between SAAR and [4] is theuse of the structured prior and semantic observations. SeeSup. Mat. for additional comparisons.

4.3. Clustering Results

We used SynthCity to obtain quantitative accuracy met-rics for the SAAR model. This allows us to study whichmodalities lead to improved semantic labeling via reduc-tion of reconstruction error. We considered various featurecombinations using appearance, orientation, location, andsemantic observations under both the TG and the MF mod-els. We quantified the utility of each of the modalities by

Page 6: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

# of geometry iterations

0 2 4 6 8 10M

ean D

ista

nce, in

ferr

ed to G

T (

m)

0.16

0.18

0.2

ToyCity

Me

an

Dis

tan

ce

, L

iDA

R-o

nly

to

GT

(m

)

0.15

0.69

Cabezas et al. CVPR14

Zhou et al. ECCV10

MF K3

MF K4

MF K5

MF K6

TG K8

TG K12

TG K16

TG K18

# of geometry iterations

0 2 4 6 8 10

Mean D

ista

nce, in

ferr

ed to G

T (

m)

0.2

0.21

0.22

0.23

ToyCity2

Me

an

Dis

tan

ce

, L

iDA

R-o

nly

to

GT

(m

)

0.2

0.73

Cabezas et al. CVPR14

Zhou et al. ECCV10

MF K3

MF K4

MF K8

MF K11

MF K12

TG K3

TG K4

TG K8

TG K11

TG K12

TG K18

TG K24

# of geometry iterations

0 2 4 6 8 10

Mean D

ista

nce, in

ferr

ed to G

T (

m)

0.57

0.58

0.59

0.6

0.61

0.62

0.63

City3

Me

an

Dis

tan

ce

, L

iDA

R-o

nly

to

GT

(m

)

0.57

0.79Cabezas et al. CVPR14

Zhou et al. ECCV10

MF K3

MF K4

TG K9

TG K12

# of geometry iterations

0 2 4 6 8 10

Mean D

ista

nce, in

ferr

ed to G

T (

m)

0.82

0.83

0.84

0.85

0.86

0.87

0.88City4

Me

an

Dis

tan

ce

, L

iDA

R-o

nly

to

GT

(m

)

0.81

1.1Cabezas et al. CVPR14

Zhou et al. ECCV10

MF K3

MF K4

TG K12

TG K14

# of geometry iterations

0 2 4 6 8 10

Mean D

ista

nce, in

ferr

ed to G

T (

m)

0.7

0.71

0.72

0.73

0.74

0.75

City1-1

Me

an

Dis

tan

ce

, L

iDA

R-o

nly

to

GT

(m

)

0.69

1.2Cabezas et al. CVPR14

Zhou et al. ECCV10

MF K3

MF K5

TG K14

TG K18

Figure 5: Mean geometry error for SynthCity reconstructions under [4] and SAAR (left y-axis); and [39] (right y-axis).

app + MF app + MF + loc app + MF + sem app + MF + loc + semapp + TG app + TG + loc app + TG + sem app + TG + loc + sem app + loc app + loc + sem

# of geometry iterations

0 1 2 3 4 5Me

an

Dis

tan

ce

, in

ferr

ed

to

GT

, (m

)

0.205

0.21

0.215

0.22

0.225

Ablation (TG-4)

# of geometry iterations

0 1 2 3 4 5Mean D

ista

nce, in

ferr

ed to G

T, (m

)

0.21

0.22

0.23

0.24

0.25

0.26

Ablation (TG-8)

# of geometry iterations

0 1 2 3 4 5Me

an

Dis

tan

ce

, in

ferr

ed

to

GT

, (m

)

0.205

0.21

0.215

0.22

0.225

Ablation (TG-11)

# of geometry iterations

0 1 2 3 4 5Me

an

Dis

tan

ce

, in

ferr

ed

to

GT

, (m

)

0.2

0.205

0.21

0.215

0.22

0.225

Ablation (TG-24)

# of geometry iterations

0 1 2 3 4 5Me

an

Dis

tan

ce

, in

ferr

ed

to

GT

, (m

)

0.205

0.21

0.215

0.22

0.225

Ablation (MF-4)

# of geometry iterations

0 1 2 3 4 5Mean D

ista

nce, in

ferr

ed to G

T, (m

)

0.21

0.22

0.23

0.24

0.25

0.26

Ablation (MF-8)

# of geometry iterations

0 1 2 3 4 5Me

an

Dis

tan

ce

, in

ferr

ed

to

GT

, (m

)

0.205

0.21

0.215

0.22

0.225

Ablation (MF-11)

# of geometry iterations

0 1 2 3 4 5Me

an

Dis

tan

ce

, in

ferr

ed

to

GT

, (m

)

0.2

0.205

0.21

0.215

0.22

0.225

Ablation (MF-24)

Figure 6: Ablation study SynthCity Toy2. Columns: Number of clusters (Ns): 4, 8, 11, 24. Rows: TG and MF models(square and diamond markers) respectively; color indicates combination of modalities: appearance, location, orientation andsemantic (see legend). Reconstruction error decreases as modalities are added, independent of model or parameters used.

measuring the error of the estimated geometry against theground-truth geometry. Fig. 6 shows that generally the re-construction accuracy improves as more features are added.For example, consider TG-24 (top-right plot): as we add thelocation feature (magenta line) to the appearance and orien-tation features (blue line) we improve reconstruction accu-racy; including the semantic feature (green line) providesfurther improvements. This behavior is seen across all TGmodels and most of the MF models. We hypothesize that asthe number of clusters grow, the MF model gets stuck in lo-cal optima and thus its performance suffers. It is importantto note that in the absence of semantic information (magentalines) the structured prior model still performs well. Fig. 7shows qualitative comparison of the labeling.

The semantic observations in SAAR can be used toattribute meaning to the learned clusters. One possiblemethod of achieving this is by analyzing the learned se-mantic component distributions ϕS . The collection of highprobability semantic observations under each cluster formsthe inferred meaning of the cluster under the model. Fig. 8

shows the semantic component distribution, MF-3 and MF-4, for the ToyCity2 scene. By looking at the learned seman-tic distribution of MF-3, we can see that the blue cluster hashigh probability observations of types: sidewalk, roads andparking lots, thus justifying the attachment of the interpre-tation of the cluster as “ground”. The labeling produced byMF-4 follows a similar pattern as MF-3. Moreover, the ef-fect of adding one more cluster component can be clearlyseen in the figure: the “ground” cluster in MF-3 (blue) isdivided into two clusters in MF-4 (blue and black). Eachof the new clusters now has a more distinct meaning: theblue cluster is solely “green space” while the black clusteris the “road network”. We note that this behavior arises nat-urally from the model and no special conditions were usedto produce this result.

4.4. Real-World Scenes

SAAR was also tested on real-world scenes. Lackingground truth, we provide qualitative comparisons. The re-sults of this section are based on the CLIF 2007 dataset [32]

Page 7: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

Figure 7: Learned categories for SynthCity City3 using TG-8 model. Colors represent cluster assignment. Left-to-Right:appearance+TG, appearance+TG+location, appearance+TG+semantic, appearance+TG+location+semantic.

Figure 8: ToyCity2 clusters. L-R: view of MF-3 and MF-4 (colors indicate cluster assignment); semantic mixture componentsϕS . Increasing Ns by one causes the ground cluster (blue MF-3) to split into ground and roads (blue and black MF-4).

Figure 9: SAAR MF-4 model. Top: Image pixels cor-responding to learned clusters (ground, buildings, roads,trees). Bottom: cluster orientations (spheres oriented so thatscene up direction points out of the page).

and the Lubbock dataset. For each of the datasets, semanticinformation was obtained from OSM [36]. Both datasetscontain semantic categories: road, building, parking lot,water, recreational and ground; additionally CLIF containsrails, while Lubbock contains grass (see Sup. Mat.). Re-constructions for these datasets are shown in Figs. 10 and11. Qualitatively these reconstructions have more detailthan the ones shown in [4]. Note that unlike prior work,SAAR clusters orientations in the correct space, i.e., the 3Dunit-sphere. This clustering is visualized in Fig. 9 for theLubbock dataset using MF-4 model. As the figure shows,orientations are indicative of scene elements; e.g., “trees”,do not have any preferred orientation. On the other hand,“ground” and “building” clusters have very compact distri-butions centered around the scene’s up direction.

4.5. Handling Missing Data

A main advantage of a probabilistic formulation is theability to easily handle noisy and missing data. Herewe show SAAR’s ability to infer cluster assignment ofpartially-observed data and predict a missing modality.Specifically, we learn primitive assignments and cluster pa-rameters using the scene’s visible data (visibility refers totriangles that have image evidence). The learned cluster pa-rameters are then used to predict the assignment of non-visible scene primitives using their location, orientation andsemantic observations. Once the non-visible primitives areassigned to a cluster, we can predict their appearance bysampling from the corresponding appearance componentposterior. The results of applying this procedure to City3and CLIF scenes are shown in Fig. 10. The smooth bound-aries between visible and non-visible assignment indicatethat the model is able to infer partial assignments well. Anexception is the river in CLIF, where the visible evidencedominate the semantic evidence for visible primitives. Dueto the small number of clusters used, four, the predicted ap-pearance captures only the main color trend.

4.6. Timing

The structured prior inference (Alg. 1) is relatively fast:approximately 0.02s per plane per iteration for TG modelsand 0.5s for MF models. This yields an overall run-timebetween 5-30 min and 30-60 min for TG and MF modelsin SynthCity respectively (workstation: 48 cores at 2.6GHzwith NVIDIA GTX 780 graphics card). Full model infer-ence (Alg. 2) is considerably slower due to the high com-

Page 8: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

Figure 10: Missing data reasoning via SAAR (MF-4). Top: CLIF; Bottom: City3. Left-to-right: visible scene primitivescolor coded according to appearance and cluster assignment, filled in cluster assignment and appearance for all primitives.

Figure 11: Lubbock reconstructions. Left: Cabezas et al. [4]; Right: SAAR (MF-4). Notice the regularized flat horizontalsurfaces obtained with SAAR.

putational costs of the geometry updates in [4]. In our non-optimized implementation, the effect of using the structuredprior, i.e., evaluating Eqs. (9) and (10), is to increase run-time by 1-3 hours. The total runtime of a geometry updatein SynthCity is between 2-20 hours (workstation: 24 coresat 2.3GHz with NVIDIA GTX Titan graphics card).

5. ConclusionWe propose SAAR, a probabilistic generative model

for inferring semantically-consistent aerial reconstructionsfrom multi-modal data. Using the novel SynthCity dataset,we have demonstrated that SAAR improves both recon-structions qualitatively and quantitatively by incorporatingsemantic data and utilizing a structured prior over geom-etry, appearance and semantic labels. Furthermore, byvirtue of the generative model construction and robust in-ference algorithm, noisy or missing data does not hinderthe model’s ability. The latter was demonstrated on two

different real-world datasets. The proposed model offersa mathematically-consistent framework for integrating bothsemantic and sensed data to generate richer scene recon-structions. Recent efforts in extending OpenStreetMap datainto the third dimension [26] can benefit from approachessimilar to the proposed one. Important extensions to thework presented here include the addition of spatial con-nectivity to the primitive’s labels; investigations of othermodalities to include for better cluster; as well as, morefine-grained categorization capabilities. All source codeas well as the novel multi-modal SynthCity dataset can bedownloaded from http://people.csail.mit.edu/rcabezas.

Acknowledgments. The authors thank Sue Zheng, Christo-pher Dean, and Oren Freifeld for general and helpful discus-sions. This research was partially supported by the Office ofNaval Research (ONR) MURI program (N000141110688)and by VITALITE, which receives support from Army Re-search Office (ARO) MURI (W911NF-11-1-0391).

Page 9: Semantically-Aware Aerial Reconstruction from Multi-Modal …people.csail.mit.edu/rcabezas/pubs/cabezas15_semantics.pdfVariables Description N P;N V;N L;N O;N C;fN I cg N C c=1 Number

References[1] S. Y. Bao, M. Bagra, Y. W. Chao, and S. Savarese. Semantic

structure from motion with points, regions, and objects. InComputer Vision and Pattern Recognition, 2012.

[2] S. Y. Bao, M. Chandraker, Y. Lin, and S. Savarese. Dense ob-ject reconstruction with semantic priors. In Computer Visionand Pattern Recognition, 2013.

[3] S. Y. Bao and S. Savarese. Semantic structure from motion.In Computer Vision and Pattern Recognition, 2011.

[4] R. Cabezas, O. Freifeld, G. Rosman, and J. W. Fisher III.Aerial Reconstructions via Probabilistic Data Fusion. InComputer Vision and Pattern Recognition, 2014.

[5] L. Cao and L. Fei-Fei. Spatially coherent latent topic modelfor concurrent segmentation and classification of objects andscenes. In ICCV, 2007.

[6] M. J. Choi, J. J. Lim, A. Torralba, and A. S. Willsky. Exploit-ing hierarchical context on a large database of object cate-gories. In Computer Vision and Pattern Recognition, 2010.

[7] P. Cignoni, C. Rocchini, and R. Scopigno. Metro: measuringerror on simplified surfaces. In Computer Graphics Forum.Wiley Online Library, 1998.

[8] L. Du, L. Ren, D. Dunson, and L. Carin. A Bayesian modelfor simultaneous image clustering, annotation and objectsegmentation. Advances in Neural Information ProcessingSystems, 2009.

[9] Esri. CityEngine. http://www.esri.com/software/cityengine.[10] N. Fioraio and L. Di Stefano. Joint detection, tracking and

mapping by semantic bundle adjustment. In Computer Visionand Pattern Recognition, 2013.

[11] A. Gelman, J. B. Carlin, H. S. Stern, and D. B. Rubin.Bayesian Data Analysis. Chapman and Hall/CRC, 2014.

[12] M. Haklay and P. Weber. Openstreetmap: User-generatedstreet maps. In Pervasive Computing, 7(4):12–18, 2008.

[13] C. Hane, C. Zach, A. Cohen, R. Angst, and M. Pollefeys.Joint 3D scene reconstruction and class segmentation. InComputer Vision and Pattern Recognition, 2013.

[14] P. Henry, M. Krainin, E. Herbst, X. Ren, and D. Fox. RGB-D mapping: Using Kinect-style depth cameras for dense 3Dmodeling of indoor environments, 2012.

[15] M. E. Hodgson and P. Bresnahan. Accuracy of AirborneLidar-Derived Elevation : Empirical Assessment and ErrorBudget. In Photogrammetric Engineering Remote Sensing,2004.

[16] B.-S. Kim, P. Kohli, and S. Savarese. 3D Scene Understand-ing by Voxel-CRF. In International Conference on ComputerVision, 2013.

[17] A. Kundu, Y. Li, F. Dellaert, F. Li, and J. M. Rehg. Joint Se-mantic Segmentation and 3D Reconstruction from Monoc-ular Video. In European Conference on Computer Vision,2014.

[18] F. Lafarge, C. Mallet. Creating large-scale city models from3D-point clouds: a robust approach with hybrid representa-tion. In IJCV, 2012.

[19] L. J. Li, R. Socher, and L. Fei-Fei. Towards total scene un-derstanding: Classification, annotation and segmentation inan automatic framework. In Computer Vision and PatternRecognition Workshops, 2009.

[20] L. J. Li, C. Wang, Y. Lim, D. M. Blei, and L. Fei-Fei. Build-ing and using a Semantivisual image hierarchy. In Confer-

ence on Computer Vision and Pattern Recognition, 2010.[21] D. Lin, S. Fidler, and R. Urtasun. Holistic scene understand-

ing for 3D object detection with RGBD cameras. In Interna-tional Conference on Computer Vision, 2013.

[22] R. Mottaghi, X. Chen, X. Liu, N.-G. Cho, S.-W. Lee, S. Fi-dler, R. Urtasun, and A. Yuille. The Role of Context forObject Detection and Semantic Segmentation in the Wild. InComputer Vision and Pattern Recognition, 2014.

[23] J.A. Nelder, R. Mead. A simplex method for function mini-mization. In The Computer Journal, 1965.

[24] Z. Niu, G. Hua, X. Gao, and Q. Tian. Spatial-DiscLDA forvisual recognition. In Computer Vision and Pattern Recog-nition, 2011.

[25] Z. Niu, G. Hua, X. Gao, and Q. Tian. Context Aware TopicModel for Scene Recognition, 2012.

[26] M. Over, A. Schilling, S. Neubauer, and A. Zipf. Gener-ating web-based 3D city models from OpenStreetMap: Thecurrent situation in Germany. Computers, Environment andUrban Systems, 34(6):496–507, 2010.

[27] X. Ren, L. Bo, and D. Fox. RGB-(D) scene labeling: Fea-tures and algorithms. In Computer Vision and Pattern Recog-nition, 2012.

[28] S. Sengupta, P. Sturgess. Semantic octree: Unifying recog-nition, reconstruction and representation via an octree con-strained higher order MRF. In ICRA, 2015.

[29] J. Straub, J. Chang, O. Freifeld, and J. W. Fisher III. ADirichlet process mixture model for spherical data. In Ar-tificial Intelligence and Statistics, 2015.

[30] J. Straub, G. Rosman, O. Freifeld, J. J. Leonard, and J. W.Fisher III. A mixture of Manhattan frames: Beyond the Man-hattan world. In CVPR, 2014.

[31] E. B. Sudderth, A. Torralba, W. T. Freeman, and A. S. Will-sky. Learning hierarchical models of scenes, objects, andparts. In ICCV, 2005.

[32] US Air Force. Columbus Large Image Format Dataset 2007.https://www.sdms.afrl.af.mil/index.php?collection=clif2007.

[33] J. P. C. Valentin, S. Sengupta, J. Warrell, A. Shahrokni, andP. H. S. Torr. Mesh based semantic modelling for indoor andoutdoor scenes. In CVPR, 2013.

[34] V. Vineet, O. Miksik, M. Lidegaard, M. Nießner,S. Golodetz, V. A. Prisacariu, O. Kahler, D. W. Murray,S. Izadi, P. Perez, P. H. S. Torr. Incremental Dense SemanticStereo Fusion for Large-Scale Semantic Scene Reconstruc-tion. In ICRA, 2015.

[35] C. Wang, D. Blei, and L. Fei-Fei. Simultaneous image clas-sification and annotation. In Computer Vision and PatternRecognition Workshops, 2009.

[36] O. Wiki. Openstreetmap wiki, 2014. [Online; accessed 18-April-2015].

[37] J. Xiao, A. Owens, and A. Torralba. SUN3D: A databaseof big spaces reconstructed using SfM and object labels. InInternational Conference on Computer Vision, 2013.

[38] O. Yakhnenko. Multi-modal hierarchical Dirichlet processmodel for predicting image annotation and image-object la-bel correspondence. SIAM Conference on Data Mining,2009.

[39] Q. Y. Zhou, U. Neumann. 2.5D Dual Contouring: A RobustApproach to Creating Building Models from Aerial LiDARPoint Clouds. In European Conference on Computer Vision,2010.


Recommended