+ All Categories
Home > Documents > Modeling urbanization patterns with generative … · MODELING URBANIZATION PATTERNS WITH...

Modeling urbanization patterns with generative … · MODELING URBANIZATION PATTERNS WITH...

Date post: 17-Sep-2018
Category:
Upload: lamphuc
View: 230 times
Download: 0 times
Share this document with a friend
4
MODELING URBANIZATION PATTERNS WITH GENERATIVE ADVERSARIAL NETWORKS Adrian Albert 1,2,* , Emanuele Strano 1,3 , Jasleen Kaur 4 , Marta Gonz ´ alez 1,2,5 1 Massachusetts Institute of Technology, Cambridge, MA 02139 2 Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA 94720 3 German Aerospace Center (DLR), Oberpfaffenhofen, Germany 4 Philips Research U.S.A., Cambridge, MA 02141 5 University of California, Berkeley CA 94720 ABSTRACT In this study we propose a new method to simulate hyper- realistic urban patterns using Generative Adversarial Net- works trained with a global urban land-use inventory. We generated a synthetic urban “universe” that qualitatively re- produces the complex spatial organization observed in global urban patterns, while being able to quantitatively recover certain key high-level urban spatial metrics. Index Termsgenerative adversarial networks, urban modeling, global urbanization 1. INTRODUCTION A long-standing question for urban and regional planners pertains to the ability to realistically simulate, by means of explicit spatial modeling, the displacement of urban land-use [1]. Modeling urban patterns has numerous applications, ranging from understanding of urbanization dynamics (which at certain scale of observation follow few physical laws [2]) to inferring future urban expansion to inform policy makers towards a better and more inclusive planning process [3]. Ur- ban models are typically classified in three main categories, land-use/transportation models, cellular automata and agent- based models [4]. These models explicitly locate the urban land-use given the interactions between spatial co-variates like location of services, population density or land price. However, data on spatial co-variates are difficult and ex- pensive to compile, and are often not available in developing countries where urban growth is more likely to occur. The recent availability of remote-sensing-based global land-use inventories and the advancements in deep learning methods offer a unique opportunity for pushing the state of the art of spatially-explicit urban models. In this study we propose a spatial explicit model of urban patterns that is based on Generative Adversarial Networks (GANs) [5] trained with very limited spatial information. GANs are a new paradigm of training machine learning models which * Corresponding author (email: [email protected]). have shown impressive results on difficult computer vision tasks such as natural images generation [6]. This is a very ac- tive area of contemporary machine learning research, whose potential to learn complex spatial distributions has only in the last year started to become better understood in the com- putational physical sciences literature. For example, recent work has leveraged GANs to generate synthetic satellite im- ages of urban environments [7, 8], de-noise telescope images of galaxies[9], or generate plausible “virtual universes” by learning from simulated data on galaxies [10]. Using a global training samples of 30, 000 cities (urban footprint scenes), we show that a basic, unconstrained GAN model is able to generate realistic urban patterns that cap- ture the great diversity of urban forms across the globe. We see this as a first step towards flexible urban land use sim- ulators for more accurate projections on urbanization in re- gions where local data is unavailable and difficult to obtain. Next, we outline the basic GAN architecture used (Sec. 2), present experimental results and an empirical validation of the model (Sec. 3). We conclude with key open questions of designing generative models for urban land use analysis (Sec. 4). All code and experiments for this study are available at https://github.com/adrianalbert/citygan. 2. MATERIALS AND METHODS 2.1. Generative adversarial networks (GANs) Generative adversarial networks (GANs) [5] represent a novel paradigm of training unsupervised machine learning models that learn representations of the input data by training two networks against each other. In the original formulation [5], a generator G receives as input a random noise vector z, which it transforms in a deterministic way (e.g., by passing it through successive deconvolutional layers if G is a deep CNN) to output a sample x fake = G(z). The discriminator D takes an input x (which can be either real, from an em- pirical dataset, or synthetically generated by G), and outputs the source probability P (o|x)= D(x) that x is either sam- arXiv:1801.02710v1 [cs.LG] 8 Jan 2018
Transcript
Page 1: Modeling urbanization patterns with generative … · MODELING URBANIZATION PATTERNS WITH GENERATIVE ADVERSARIAL NETWORKS Adrian Albert 1 ;2 ;, Emanuele Strano 1 ;3, Jasleen Kaur

MODELING URBANIZATION PATTERNS WITH GENERATIVE ADVERSARIALNETWORKS

Adrian Albert1,2,∗ , Emanuele Strano1,3, Jasleen Kaur4, Marta Gonzalez1,2,5

1Massachusetts Institute of Technology, Cambridge, MA 021392Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA 94720

3German Aerospace Center (DLR), Oberpfaffenhofen, Germany4Philips Research U.S.A., Cambridge, MA 02141

5University of California, Berkeley CA 94720

ABSTRACT

In this study we propose a new method to simulate hyper-realistic urban patterns using Generative Adversarial Net-works trained with a global urban land-use inventory. Wegenerated a synthetic urban “universe” that qualitatively re-produces the complex spatial organization observed in globalurban patterns, while being able to quantitatively recovercertain key high-level urban spatial metrics.

Index Terms— generative adversarial networks, urbanmodeling, global urbanization

1. INTRODUCTION

A long-standing question for urban and regional plannerspertains to the ability to realistically simulate, by means ofexplicit spatial modeling, the displacement of urban land-use[1]. Modeling urban patterns has numerous applications,ranging from understanding of urbanization dynamics (whichat certain scale of observation follow few physical laws [2])to inferring future urban expansion to inform policy makerstowards a better and more inclusive planning process [3]. Ur-ban models are typically classified in three main categories,land-use/transportation models, cellular automata and agent-based models [4]. These models explicitly locate the urbanland-use given the interactions between spatial co-variateslike location of services, population density or land price.

However, data on spatial co-variates are difficult and ex-pensive to compile, and are often not available in developingcountries where urban growth is more likely to occur.

The recent availability of remote-sensing-based globalland-use inventories and the advancements in deep learningmethods offer a unique opportunity for pushing the stateof the art of spatially-explicit urban models. In this studywe propose a spatial explicit model of urban patterns thatis based on Generative Adversarial Networks (GANs) [5]trained with very limited spatial information. GANs are anew paradigm of training machine learning models which

∗Corresponding author (email: [email protected]).

have shown impressive results on difficult computer visiontasks such as natural images generation [6]. This is a very ac-tive area of contemporary machine learning research, whosepotential to learn complex spatial distributions has only inthe last year started to become better understood in the com-putational physical sciences literature. For example, recentwork has leveraged GANs to generate synthetic satellite im-ages of urban environments [7, 8], de-noise telescope imagesof galaxies[9], or generate plausible “virtual universes” bylearning from simulated data on galaxies [10].

Using a global training samples of 30, 000 cities (urbanfootprint scenes), we show that a basic, unconstrained GANmodel is able to generate realistic urban patterns that cap-ture the great diversity of urban forms across the globe. Wesee this as a first step towards flexible urban land use sim-ulators for more accurate projections on urbanization in re-gions where local data is unavailable and difficult to obtain.Next, we outline the basic GAN architecture used (Sec. 2),present experimental results and an empirical validation ofthe model (Sec. 3). We conclude with key open questions ofdesigning generative models for urban land use analysis (Sec.4). All code and experiments for this study are available athttps://github.com/adrianalbert/citygan.

2. MATERIALS AND METHODS

2.1. Generative adversarial networks (GANs)Generative adversarial networks (GANs) [5] represent a novelparadigm of training unsupervised machine learning modelsthat learn representations of the input data by training twonetworks against each other. In the original formulation [5],a generator G receives as input a random noise vector z,which it transforms in a deterministic way (e.g., by passingit through successive deconvolutional layers if G is a deepCNN) to output a sample xfake = G(z). The discriminatorD takes an input x (which can be either real, from an em-pirical dataset, or synthetically generated by G), and outputsthe source probability P (o|x) = D(x) that x is either sam-

arX

iv:1

801.

0271

0v1

[cs

.LG

] 8

Jan

201

8

Page 2: Modeling urbanization patterns with generative … · MODELING URBANIZATION PATTERNS WITH GENERATIVE ADVERSARIAL NETWORKS Adrian Albert 1 ;2 ;, Emanuele Strano 1 ;3, Jasleen Kaur

G DCo

nvTr

ansp

Batc

hNor

m

Tanh

ReLU

G

a

b

Fig. 1. a) GAN model architecture following [5]; b) the archi-tecture for generator G following [6] is composed of inverse-convolutional, batch normalization, and rectified linear unit(ReLU) layers (the architecture for D is similar).

pled from the real distribution (o = real), or produced by G(o = fake). When G is optimal, xfake is implicitly sampledfrom the data distribution that G seeks to emulate. This pro-cess is summarized schematically in Figure 3a).

Both G and D are deep convolutional neural networksparametrized by the weights vectors θG and θD. Theseweights are learned via back-propagation [5] by alternativelyminimizing the following loss functions:

θD :LD = Ex∼px[log D(x)] + Ez∼pz

[log(1−D(G(z)))],(1)

θG :LG = Ez∼pz[log (1−D(G(z)))] (2)

The architectures we used for both G and D followclosely those proposed in [6]. The generator architecture is il-lustrated in Figure 3b). It is composed of is composed of sev-eral convolutional blocks consisting of inverse-convolutional,batch normalization, and rectified linear unit (ReLU) layers,ending in a hyperbolic tangent layer (which applies a tanh(·)nonlinearity to each element of the generated map). For thediscriminator D we used a very similar architecture, withthe only differences being leaky ReLU non-linearities insteadof the ReLU non-linearities in G and convolutional layersinstead of transposed convolutions.

2.2. Training sample: built-up areas at global-scale

Here we focused on the simplest, and arguably the most in-formative spatial feature of cities, which is the presence ofbuilt-up areas. To construct a training sample, we used the“Global Urban Footprint” (GUF)[11], an updated global in-ventory of built-up land at∼ 12m/px resolution. This datasetis published by the German Aerospace Center (DLR) and

has been obtained through extensive processing of syntheticaperture radar (SAR) satellite scenes acquired between 2011-2012. We used the built-up footprint of all cities with at least10, 000 inhabitants (which we estimated by combining pop-ulation estimates from the LandScan data [12] with city ad-ministrative boundaries worldwide from the GADM dataset[13]). For each city, we extracted a sample maps as a squarewindows of 100 × 100km centered on city center. Fixing aspatial scale ofL = 100km results in different image sizes (inpixels) for cities at different latitudes on Earth. We aggregateeach extracted map at 750m/px resized to 128× 128 pixels.The final training dataset contains N = 29, 980 binary maps(images) xi, i = 1, ..., N , with xi ∈ RW×W and W = 128.

3. MODEL RESULTS AND VALIDATION

Having trained a generator G, we simulated a synthetic “ur-ban universe” of 30, 000 urban maps. Figure 2 illustratesrandomly-selected real (left panel) with simulated urban pat-terns (right panel). At a visual inspection simulations arepractically indistinguishable from the real scenes, exhibitingrealistic concentrations and spreads of urban masses, includ-ing those characteristic of coastal or inland cities. This is inthe absence of externally-imposed constraints (e.g., inform-ing the model that water areas cannot be built up). However,aside from qualitative comparisons, it is difficult to quantifythe “realism” of a simulated city, since humans have not in-nate abilities to recognize remote-sensing images of cities (asit is the case in natural images that gave rise to metrics like“Inception score” to quantify GAN performance [14]).

Thus, our validation strategy is to use spatial summarystatistics on urban form to compare real against simulatecities. The average radial profile x(d) [15] is perhaps oneof the most widely-accepted such tools in the urban anal-ysis literature. We compute x(d) by averaging the totalamount of built-up area x within rings of width ∆d at at adistance d from the center (see Figure 3, i.e., values x(u, v) :(u, v) ∈ R(d), withR(d) ≡ {(u, v)|(u−u0)2+(v−v0)2 >d2 and (u− u0)2 + (v − v0)2 ≤ (d+ ∆d)2}:

x(d) ≡ 1

|R(d)|∑

(u,v)∈R(d)

x(u, v) (3)

We used the radial profiles in Eq. (3) to determine thepolycentric nature of real and simulated scenes via a peak-search algorithm, as illustrated in Fig.3. The peak search al-gorithm finds points in a univariate profile whose value (peakheight) as fraction of maximum is at least h, and at a distancefrom a previously-identified peak of at least δ. We set accept-able values h = 50% and δ = 5 km via experimentation.In Figure 4a) we compare the distributions of the number ofpeaks for all cities on Earth with that for the simulated ur-ban universe. The two distributions show similar form (thep-value on a χ2 test is ∼ 10−6).

As a further validation, we clustered the radial profiles ofreal cities and compared to the typical profiles of synthetic

Page 3: Modeling urbanization patterns with generative … · MODELING URBANIZATION PATTERNS WITH GENERATIVE ADVERSARIAL NETWORKS Adrian Albert 1 ;2 ;, Emanuele Strano 1 ;3, Jasleen Kaur

cihanbeyli,turkeypop:101.4k

praha--zpad,czech-republicpop:136.9k

wiener-neustadt,austriapop:41.1k

terrabona,nicaraguapop:15.7k

panna,indiapop:922.3k

praha-19,czech-republicpop:38.1k

praha--vchod,czech-republicpop:138.5k

faisalabad,pakistanpop:14.0M

ueda,japanpop:168.0k

tessala-lamtai,algeriapop:19.5k

wyong--south-and-west,australiapop:93.1k

central-baddibu,gambiapop:19.0k

pennington,united-statespop:12.5k

igembe-south,kenyapop:32.8k

ouled-sellem,algeriapop:23.2k

bodinga,nigeriapop:220.5k

paran,argentinapop:295.2k

shoalhaven--pt-a,australiapop:34.0k

bargarh,indiapop:1.1M

thaphalanxay,laospop:37.9k

maynskiy-rayon,russiapop:10.7k

ramsey,united-statespop:589.3k

miguel-alves,brazilpop:21.2k

czuchw,polandpop:37.0k

Fig. 2. Comparing real urban built land use maps (left) with synthetic maps (right) simulated with a Generative AdversarialNetwork (GAN). In each case the pixel values are in [0, 1] and represent the fraction of land occupied by buildings.

Paris Mumbai Rio de Janeiro

Fig. 3. Built land use maps of three example cities (upperrow) with their average radial profiles (bottom row). The reddots indicate the peaks detected by the peak-search algorithm.

ones. To cluster the profiles, we used the K-Means algo-rithm [16]. The results are summarized in Fig. 4. Usinga simple fraction of sum-of-squares argument [16], we iden-tify K = 12 as the best number of clusters for both real andsynthetic scenes. As shown in Fig. 4 b) the profile classes aregenerally very similar as also visible in the panel c) where dis-tribution of number of scenes per classes is shown. The dis-tributions are again similar, although for the classes 3 (mono-centric cities) and 8 (sprawled patterns) we observe larger dif-ferences. We argue that such differences can be due of thesampling strategy which would have favor the abundance ofmono-centric urban patterns, while the simulation have beengenerated regardless the position of the urban core. Note that,by computing average centroids for each of the profile classes,narrower peaks get averaged out. This is an artifact of “mea-suring” spatial built land use maps in this simple way; indeed,the peak (layer) count and average profile class offer two com-plementary views on which to compare spatial distributions.

4. DISCUSSION AND CONCLUSIONS

In this study we shown, for the first time, that modern gen-erative machine learning models such as GANs can success-fully be used to simulate realistic urban patterns. This is buta start, and despite the impressive results important severalopen questions still remain. Most of them, as typically fordeep-learning (DL) models, pertain to the black-box natureof deep neural networks, which currently lack comprehen-sive human interpretability and ability for fine-tuned control.We believe, however, that this limitation, which certainly de-serves (and gets) attention in the DL literature (e.g., [17])should not preclude research into their promise to augmentexisting models using globally-available remote-sensing data.

Important open questions remain: How to evaluate thequality of model output in a way that is both quantitative, andinterpretable and intuitive for urban planning analysis? Howto best disentangle, explore, and control latent space repre-sentations of important characteristics of urban spatial maps?How to learn from both observational and simulated data oncities? In addition, this initial work has only focused on astatic snapshot; another area of research is to model city evo-lution over time using available remote-sensing data (e.g., viathe GUF dataset on built land we used here). We plan to ad-dress these open questions in on-going work.

5. REFERENCES

[1] Michael Batty, “Model cities,” Town Planning Review,vol. 78, no. 2, pp. 125–151, 2007.

[2] Michael Batty and Paul A Longley, Fractal cities: ageometry of form and function, Academic press, 1994.

[3] John D Landis, “Urban growth models: State of the

Page 4: Modeling urbanization patterns with generative … · MODELING URBANIZATION PATTERNS WITH GENERATIVE ADVERSARIAL NETWORKS Adrian Albert 1 ;2 ;, Emanuele Strano 1 ;3, Jasleen Kaur

# peaks

Pro�le Class (C)

C0 C1 C2 C3

C4 C5 C6 C7

C8 C9 C10 C11

global samplesimulations

a b

c

Fig. 4. a): the distribution of number of satellite urban centers for real and synthetic cities; c): the distribution of radial profileclasses for real and synthetic cities; b): typical radial profiles for real and synthetic cities.

art and prospects,” Global urbanization, pp. 126–140,2011.

[4] Xuecao Li and Peng Gong, “Urban growth models:progress and perspective,” Science Bulletin, vol. 61, no.21, pp. 1637–1650, 2016.

[5] I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu,D. Warde-Farley, S. Ozair, A. Courville, and Y. Ben-gio, “Generative Adversarial Networks,” ArXiv e-prints,June 2014.

[6] Alec Radford, Luke Metz, and Soumith Chintala, “Un-supervised representation learning with deep convolu-tional generative adversarial networks,” CoRR, vol.abs/1511.06434, 2015.

[7] DaoYu Lin, “Deep unsupervised representationlearning for remote sensing images,” CoRR, vol.abs/1612.08879, 2016.

[8] Nikolay Jetchev, Urs Bergmann, and Roland Vollgraf,“Texture synthesis with spatial generative adversarialnetworks,” CoRR, vol. abs/1611.08207, 2016.

[9] K. Schawinski, C. Zhang, H. Zhang, L. Fowler, andG. K. Santhanam, “Generative adversarial networks re-cover features in astrophysical images of galaxies be-yond the deconvolution limit,” Monthly notices of theroyal astronomical society, vol. 467, pp. 110–114, May2017.

[10] M. Mustafa, D. Bard, W. Bhimji, R. Al-Rfou, andZ. Lukic, “Creating Virtual Universes Using GenerativeAdversarial Networks,” ArXiv e-prints, June 2017.

[11] Thomas Esch, Wieke Heldens, Andreas Hirner, Man-fred Keil, Mattia Marconcini, Achim Roth, Julian Zei-dler, Stefan Dech, and Emanuele Strano, “Breaking newground in mapping human settlements from space theglobal urban footprint,” ISPRS Journal of Photogram-metry and Remote Sensing, vol. 134, no. Supplement C,pp. 30 – 42, 2017.

[12] Oak Ridge National Laboratory, “Landscan global pop-ulation dataset 2013,” Oak Ridge, Tennessee, 2014.

[13] Robert Hijmans, Julian Kapoor, John Wieczorek, NelGarcia, Aileen Maunahan, Arnel Rala, and Alex Man-del, “Global administrative areas (gadm),” Univer-sity of California, Davis. Available online at http://www.gadm.org/version2, 2016.

[14] Tim Salimans, Ian J. Goodfellow, Wojciech Zaremba,Vicki Cheung, Alec Radford, and Xi Chen, “Im-proved techniques for training gans,” CoRR, vol.abs/1606.03498, 2016.

[15] Alain Bertaud and Stephen Malpezzi, “The spatial dis-tribution of population in 48 world cities: implicationsfor economies in transition,” Technical report, 2003.

[16] Trevor Hastie, Robert Tibshirani, and Jerome Friedman,The Elements of Statistical Learning, Springer Seriesin Statistics. Springer New York Inc., New York, NY,USA, 2001.

[17] Mehdi Mirza and Simon Osindero, “Conditional gener-ative adversarial nets,” CoRR, vol. abs/1411.1784, 2014.


Recommended