-
Notifications
You must be signed in to change notification settings - Fork 0
/
s1_intro.tex
286 lines (266 loc) · 19.3 KB
/
s1_intro.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
\section{Introduction}
\label{sec:intro} % ca. 1500 words
% Urban form and function % Important to understand
The way in which different urban functions are arranged within space, and the forms
these give rise to, are important to understand how cities work, how they interact with
the human and environmental systems that create them, and how policy can effectively
intervene.
%% Why? Encodes history, conditions the future
Urban form and function matter, at least, for two reasons \citep{dab_mf_2021a}: first
because cities use both to encode their history; and second because, once in place, the
physical layout of functions within a city condition how it can and will develop in the
future.
%% Key to understanding is measurement
A key requirement to understand form and function in cities is adequate measurement,
which implies detailed, consistent, and scalable characterisations that can be updated
frequently over time. These characteristics then allow not only to observe detail, but
to see it unfold both over space and time.
%% Rare and incomplete currently, of detailed, scalable and consistent, pick any two
There is a large literature measuring these phenomena, and it is relatively common to
find any two of those characteristics (i.e., detailed and consistent, consistent and
scalable, and detailed and scalable) present in a given piece of work.
%% Recent developments are changing this --> spatial signatures
Research bringing the three together is still rare, although some is emerging (e.g.,
\citealp{fleischmann2022geographical}) thanks to the confluence of better data, open
source software, and cheap computing power.
%% Detailed measurement is expensive in terms of time, effort, and data requirements
Still, generating detailed, consistent, and scalable classifications of urban form and
function is an expensive process that is difficult to refresh regularly because most of
the underlying data sources only see updates infrequently.
% The promise of satellites % A promising solution is to _supplement_ detailed
%measurements with satellite imagery
A promising option to improve the frequency of these classifications is satellite
imagery.
%% Satellite has radically improved in the last years, and it is set to continue on that
%technological path
Satellite technology has radically increased and improved the amount of data available
on the Earth, and shows no signs of slowing down.
%% At the same time, the algorithms to process satellite have also seen a revolution in
%the last ten years (deep learning)
More and better imagery has been complemented with the rise of new computer vision
algorithms, such as deep learning \citep{lecun2015}, that allow to extract more value
from the same amount of data; and the availability of computing power that makes it
possible to deploy them cheaply without the steep learning curve required only a few
years ago.
%% These two trends converge in making possible things with satellite that was
%unthinkable a few years ago
The convergence of these three trends in remote sensing is unlocking achievements that
even very recently seemed beyond the realm of possibility.
%% Such as measuring UFF using and end-to-end open pipeline (data and code)
One such area is the use of remote sensing and satellite technology to decode complex
patterns in urban landscapes, such as the spatial signature of different types of form
and function.
%% Open is important because it multiplies the options of what is possible to do with
%outputs
Just as importantly, many of these advances are being built atop technology developed
under open licenses that allow to further build on them, freely redistributing
downstream outputs.
% Lit. review %
% Satellites for cities
The use of satellite technology for measuring different aspects of urban environments is
by no means new.
%% Mostly through Remote Urban Sensing
Much of the present work falls within the broad category of urban remote sensing
\citep{rashed2010remote, weng2018urban, yang2021urban}. In fact, the promise of using
remote sensing data to decode the complexity of urban structure has long been recognised
(e.g., \citealp{longley2002geographical}).
%% The vast majority is as supervised object detection --> include building footprints
Much of the work in this area has traditionally focused on identification of individual
geographic features, such as building footprints (e.g., \citealp{microsoft2019}) or
trees (e.g., \citealp{ke2011review}). More recently, the field has started to pay
increasing attention to the use of modern algorithms such as deep learning
\citep{lai2021deep}, and attempting to map more complex patterns that involve bundles of
features rather than a single one (e.g., \citealp{kuffer2021mapping}).
%% And as Land Use and Land Cover --> Include recent examples (ESRI land cover, Google
%Dynamic World, etc.)
On the adjacent domain of Land Use / Land Cover (LULC) mapping, recent advances have
shown the potential of using frequently updated, open satellite data in combination with
modern computer vision to effectively map land cover globally in quasi continuous ways
(e.g., \citealp{karra2021global, brown2022dynamic}; see \citealp{venter2022global} for a
detailed comparison of some of the most novel data products in this realm).
% Satellites for urban form and function % Much less on recognising composite patterns
%(e.g., UFF) instead of single objects or uses
While most of the efforts in urban remote sensing have focused on the identification of
individual features or single uses, much less work has been directed at decoding
patterns that involve several features and/or uses to be identified.
%% In some ways, this is more complicated but, in others, maybe not (plus we have much
%better tools now!)
In some ways, the jump from the simpler goal of identifying one object or a single use
to detecting a pattern that involves a particular bundle of them is not without its
challenges and shortcomings \citep{wang2022knowledge}. But, given the performance of
modern algorithms, and the increase in resolution and quality of even openly available
imagery, realising this goal is starting to become possible.
%% Much focused on Local Climate Zones and slums
There are two areas that have received most of the attention in this context. One
revolves around the prediction of Local Climate Zones (LCZs, \citealp{stewart2012}).
LCZs are a set of pre-defined classes of urban fabric originally developed for the study
of the urban heat island effect. A growing body of literature has focused on developing
more exhaustive and sophisticated models to extract these classes from satellite imagery
(e.g., \citealp{koc2017mapping, wang2018mapping, liu2020local, taubenbock2020,
zhou2021parcel, zhou2022deep}).
% Slums
The second one is focused on one particular type of urban form and function that is
mostly found in regions which are typically data scarce: informal settlements, or urban
slums. For the interested reader, \cite{slums2016} provides an excellent starting point.
% A bit of morphometrics
Although much more in its infancy, a nascent area of interest is growing around using
imagery to decode urban form (e.g., \citealp{04f9ab8f6c714010ac39b58230f59d85}).
% % Deep Learning Architectures
% A common element of the recent advances reviewed above is the use of deep convolutional
% neural networks to perform the task of interest (i.e.,
% classification/segmentation/recognition) from satellite imagery.
% % From scratch
% Some studies, particularly those with sufficient data and computation available, train
% networks from scratch. These involves sometimes building a bespoke architecture (e.g.,
% \citealp{othman2017domain}), bespoke data for training (e.g., \citealp{qiu2020fusing,
% karra2021global}), or both (e.g., \citealp{taubenbock2020, zhu2022urban,
% sharma2017patch, wang2018multi}).
% % Transfer learning
% Other works however rely on existing architectures like VGG/16 \citep{simonyan2014very},
% UNet \citep{ronneberger2015u}, or ResNet \citep{he2016deep}; standardised databases such
% as ImageNet \citep{ILSVRC15} for training; or both (e.g., \citealp{qiu2020fusing,
% karra2021global, srivastava2019understanding}). The letter is known as \textit{transfer
% learning} and usually involves re-training of the top layers of the network to customise
% predictions to the specific use case, while retaining unchanged the original weights for
% all the other layers.
%% List from MF on examples %
%https://github.com/urbangrammarai/signature_ai_paper/issues/5
%%%%%%%%%%%%%%%
% RS focusing on urban env are of three kinds
% LCLZ whose focus is not primarily on urban and only a fraction of their classes
% are urban Built/unbuilt type of classifications (include JRC with 4 levels) LCZ
% with pre-defined classes with an RS application in mind
% The understanding on how much of urban insight can we see from above is still very
% limited This paper takes a classification that 1) flips the ratio of urban/non-urban
% classes compared to LCLZ; 2) is data driven, with no particular secondary use case in
% mind (i.e. not designed to be seen); and asks the following questions:
% can we reliably detect this type and granularity of urban classification from open
% satellite imagery? given the geographical nature of the taks, is there a scope for
% spatially-explicit methods built on top of whatever is neural net doing? would
% this potentially allow us to use AI and ML to potentially rollout time series of
% signatures based on Sentinel 2?
A common element of the recent advances reviewed above is the use of deep convolutional
neural networks to perform the task of interest (i.e.,
classification/segmentation/recognition) from satellite imagery. While neural networks
are becoming ubiquitous in the analysis of urban satellite imagery, their application has
so far mostly ignored the geographical nature of the images being fed to these
algorithms. This is not entirely unreasonable. Much of the state of the art in deep
learning and computer vision was developed in the last decade with ``aspatial imagery''
in mind, in particular consumer photographs uploaded and shared through the internet
(e.g., featuring cats and dogs). As such, many of the assumptions (e.g., unrelated
images), tricks (e.g., data augmentation techniques), and limitations (e.g., shape of
the input data) these models feature are intimately related to data of this kind. The
application of deep learning to satellite imagery is in what we consider a first phase
in which cutting-edge computer vision has been deployed to images that, rather than
animals or people, represent locations on Earth observed from above. Because of the
overall performance of modern algorithms, the results are impressive, even
with largely unmodified models. However, this does not imply there is no further margin
for improvement.
The main aspect of the geographical nature is that the individual images sampled from
the satellite imagery are not independent from each other, but rather are part of a
continuous whole that is the Earth's surface. This means that the spatial configuration
of the images is as important as the content of each image. This is a key difference
with the standard computer vision tasks that have been the focus of most of the research
in the field so far. By ignoring this aspect, we are leaving value on the table when
analysing satellite imagery. The architecture of convolutional neural networks is able
to capture some of the spatial information but only within the confines of the
individual image. Looking at the spatial statistics methods, we learn that the inclusion
of geographical context in the model, often in the form of a spatially smoothed average
(sometimes referred to as a ``spatial lag''), is a core component and one of the key
distinctions between statistics and spatial statistics. We believe that such a
distinction should happen in the realm of computer vision as well and that the
geographical nature of the data should play an explicit role in the design of the
algorithms.
The explicit spatial dimension can be embedded in many ways. The first one is to alter
the architecture of the neural network to include not only the content of the image but
also its neighborhood as an input. In practice, two chips would be fed to the network:
given the image of 16x16 pixels, the network would also receive, e.g., a 32x32 image,
with the original image in the center and the surrounding pixels as the context. The
second one is to use the output of the neural network as an input to a spatial model
that would smooth the predictions based on the spatial configuration of the images. Both
are equally valid but have different implications in terms of computational complexity
and interpretability. The first one is more computationally demanding as it requires the
network to process two images at once, making the result less interpretable at the same
time as we would not be able to tell which part of the input is driving the result the
most. The second one is less computationally demanding while allowing for different
models to be used for the final prediction based on the spatial configuration. Given
these models can be anything from logistic regression to gradient-boosted trees, the
final model allows for analysis of the importance of spatial configuration allowing for
a more nuanced interpretation. Both approaches provide a way to incorporate inherent
spatial autocorrelation of the data in the model, making explicit use of Tobler's First
Law of Geography \citep{tobler1970computer}. Another approach is to use the geographical
nature of the data in spatial augmentation, allowing for sampling using a ''sliding
window''. All further require a consideration of the image size sampled from continuous
satellite data as the number of pixels directly affects the scale and inherent spatial
unit of the analysis, leading to issues known as Modifiable Areal Unit Problem (MAUP)
\citep{openshaw1981modifiable}. At the same time, all require a careful design of the
experiments to ensure that the spatial context is not leaking from the training to the
validation set.
This paper focuses on a subset of these options. We test the role of the scale of the
image, the effect of spatial augmentation using the sliding technique, and the role of a
modelling including spatial context on top of the neural network output. It starts from
an existing classification of Great Britain that is data-driven, designed to best
capture urban form and function from available data (i.e. it is not designed to be seen
on satellite imagery), and that flips the ratio of urban vs non-urban classes compared
to most LULC classifications. From there, we build a matrix of experiments that allow us
to test 1) the scale of the input image, 2) the effect of spatial augmentation, and 3)
the role of modelling on top of neural network outputs and inclusion of spatial context
in the final prediction. The key methodological advancement of this paper lies in the
latter, which also proves to be the only consistent way to improve the predictions.
The remainder of the paper is structured as follows: Section \ref{sec:matmet} describes
the data, covering spatial signatures as the target of the prediction and Sentinel 2
satellite imagery as the data used to predict signatures as well as the methodological
strategy we follow, reflecting all chip size selection, spatial data augmentation, model
architecture, performance metrics, and a method of experiment summarization; Section
\ref{sec:results} presents the key results from our experiments in a form of tables and
figures; and Section \ref{sec:discussion} discusses the relevance of each of the
dimensions of the experiment matrix, the performance of the models, and the implications
of the results for the design of spatially explicit methods within remote sensing.
% With many existing methods, there is a tendency towards a simplification of urban
% environment. In the most simple cases, we can talk about differentiation between
% built-up areas and non-built-up land, occasionally expanded to a few gradual steps in
% between (e.g. degree of urbanization by \cite{DIJKSTRA2021103312}). Land use and land
% cover classifications focus primarily on non-urban landscape, leaving all patterns
% that make up cities in a handful of classes. LCZ classification is based on
% pre-defined conceptual classes designed with an RS application in mind, omitting
% distinctions that are not easy to see on imagery. The understanding on how many types
% of urban insight can be seen from above is still very much an unexplored area.
% This paper advances our understanding of the extent to which deep learning methods can
% be applied to satellite imagery in order to capture the composition of primarily
% urbanized landscapes. It starts from an existing classification of Great Britain that
% is data-driven, designed to best capture urban form and function from available data
% (i.e. it is not designed to be seen on satellite imagery), and that flips the ratio of
% urban vs non-urban classes compared to most LULC classifications. From there, we move
% towards answering whether we can reliably detect this type and granularity of urban
% classification from open satellite imagery. Furthermore, given the geographical nature
% of the task, whether the is scope for spatially-explicit methods built on top of the
% output of neural network to bridge the gap between computer vision and geography?
% Understanding these topics can potentially allow development of a time series of
% otherwise static classifications like the one used in this paper and uncover the
% evolution of urban form and function in cities. The remainder of the paper is
% structured as follows: Section \ref{sec:matmet} describes the data we use as well as
% the methodological strategy we follow; Section \ref{sec:results} presents the key
% results from our experiments; and Section \ref{sec:discussion} discusses their
% relevance and concludes.
%% Much of the research above is focused on directly deploying standard computer imagery
%algorithms % The aspect of Geography is largely ignored % What do we mean by "the
%aspect of Geography"? %% How to treat images that represent geographical locations %%
%Some key differences with traditional computer vision: %% - images are "arbitrarily"
%cut from a continuous one --> role of scale %% - images are also thus intrinsically
%connected to each other through their spatial configuration --> role of context %
%Ignoring this means we are leaving "value on the table" when analysing satellite
%imagery % and standard computer vision does not provide off-the-shelf approaches
% This paper % Focus on the geographical aspect to explore its role in identifying UFF
%from satellites % We use state of the art AI/computer vision, but only as the starting
%point to explore geography % We set up a set of experiments that allow us to test a
%series of hypotheses on the role of scale and context Results % Scale and context
%matter, and these insights can be incorporated to improve predictions The remainder of
%this paper is structured as follows
%------------------------------------------------------------------------------------
%X Keep it conceptual about the point of the paper
%X Include literature review X (focused on what is available at the intersection of
%satellite + AI)
%X Highlight what the key missing bits are when AI is applied to spatial X imagery
%(e.g., scale and context)
%X a lot of the ai and satellite has been supervised object detection X there is not a
%lot of ai for patterns rather than features - in some way it may be easier