Object-based classification of remote sensing data for change detection Volker Walter * Institute for Photogrammetry, University of Stuttgart, Geschwister-Scholl-Str. 24 D, Stuttgart D-70174, Germany Received 31 January 2003; accepted 26 September 2003 Abstract In this paper, a change detection approach based on an object-based classification of remote sensing data is introduced. The approach classifies not single pixels but groups of pixels that represent already existing objects in a GIS database. The approach is based on a supervised maximum likelihood classification. The multispectral bands grouped by objects and very different measures that can be derived from multispectral bands represent the n-dimensional feature space for the classification. The training areas are derived automatically from the geographical information system (GIS) database. After an introduction into the general approach, different input channels for the classification are defined and discussed. The results of a test on two test areas are presented. Afterwards, further measures, which can improve the result of the classification and enable the distinction between more land-use classes than with the introduced approach, are presented. D 2003 Elsevier B.V. All rights reserved. Keywords: change detection; classification; object-oriented image analysis; data fusion 1. Introduction In Walter and Fritsch (2000), a concept for the automatic revision of geographical information sys- tem (GIS) databases using multispectral remote sens- ing data was introduced. This approach can be subdivided into two steps (see Fig. 1). In a first step, remote sensing data are classified with a supervised maximum likelihood classification into different land- use classes. The training areas are derived from an already existing GIS database in order to avoid the time-consuming task of manual acquisition. This can be done if it is assumed that the number of changes in the real world is very small compared with the number of all GIS objects in the database. This assumption is justified because we want to realise update cycles in the range of several months. In a second step, the classified remote sensing data have to be matched with the existing GIS objects in order to find those objects where a change occurred, or which were collected wrongly. We solved this task by measuring per object the percentage, homogeneity, and form of the pixels, which are classified to the same object class as the respective object stored in the database (Walter, 2000). All objects are classified into the classes fully verified, partly verified, and not found by using thresholds that can be defined interactively by the user. The problem of using thresholds is that they are data-dependent. For example, the percentage of veg- 0924-2716/$ - see front matter D 2003 Elsevier B.V. All rights reserved. doi:10.1016/j.isprsjprs.2003.09.007 * Tel.: +49-711-121-4091; fax: +49-711-121-3297. E-mail address: [email protected] (V. Walter). www.elsevier.com/locate/isprsjprs ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225 – 238
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
www.elsevier.com/locate/isprsjprs
ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238
Object-based classification of remote sensing data
for change detection
Volker Walter*
Institute for Photogrammetry, University of Stuttgart, Geschwister-Scholl-Str. 24 D, Stuttgart D-70174, Germany
Received 31 January 2003; accepted 26 September 2003
Abstract
In this paper, a change detection approach based on an object-based classification of remote sensing data is introduced. The
approach classifies not single pixels but groups of pixels that represent already existing objects in a GIS database. The approach
is based on a supervised maximum likelihood classification. The multispectral bands grouped by objects and very different
measures that can be derived from multispectral bands represent the n-dimensional feature space for the classification. The
training areas are derived automatically from the geographical information system (GIS) database.
After an introduction into the general approach, different input channels for the classification are defined and discussed. The
results of a test on two test areas are presented. Afterwards, further measures, which can improve the result of the classification
and enable the distinction between more land-use classes than with the introduced approach, are presented.
D 2003 Elsevier B.V. All rights reserved.
Keywords: change detection; classification; object-oriented image analysis; data fusion
1. Introduction the real world is very small compared with the number
In Walter and Fritsch (2000), a concept for the
automatic revision of geographical information sys-
tem (GIS) databases using multispectral remote sens-
ing data was introduced. This approach can be
subdivided into two steps (see Fig. 1). In a first step,
remote sensing data are classified with a supervised
maximum likelihood classification into different land-
use classes. The training areas are derived from an
already existing GIS database in order to avoid the
time-consuming task of manual acquisition. This can
be done if it is assumed that the number of changes in
0924-2716/$ - see front matter D 2003 Elsevier B.V. All rights reserved.
of all GIS objects in the database. This assumption is
justified because we want to realise update cycles in
the range of several months.
In a second step, the classified remote sensing data
have to be matched with the existing GIS objects in
order to find those objects where a change occurred, or
which were collected wrongly. We solved this task by
measuring per object the percentage, homogeneity, and
form of the pixels, which are classified to the same
object class as the respective object stored in the
database (Walter, 2000). All objects are classified into
the classes fully verified, partly verified, and not found
by using thresholds that can be defined interactively by
the user.
The problem of using thresholds is that they are
data-dependent. For example, the percentage of veg-
Fig. 1. Pixel-based classification approach.
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238226
etation pixels varies significantly between data that
are captured in summer or in winter. Other influencing
factors are light and weather conditions, soil type, or
daytime. Therefore, we cannot use the same thresh-
olds for different datasets. In order to avoid the
problem of defining data-dependent thresholds, we
introduce an object-based supervised classification
approach. The object-based classification works in
the same way as a pixel-based classification (see
Fig. 2), with the difference that we do not classify
each pixel but combine all pixels of each object and
classify them together. Again, the training areas for
the classification of the objects are derived from the
existing database in order to avoid a time-consuming
manual acquisition.
In a ‘‘normal’’ classification, the greyscale values
of each pixel in different multispectral channels and
possibly some other preprocessed texture channels are
used as input. For the classification of groups of
pixels, we have to define new measures that can be
very simple (e.g., the mean grey value of all pixels of
an object in a specific channel) but also very complex,
like measures that describe the form of an object. This
approach is very flexible because it can combine very
different measures for describing an object. We can
even use the result of a pixel-based classification and
count for each object the percentage of pixels that are
classified to a specific land-use class.
Because the result of the approach is a classifica-
tion into the most likely class, the problematic part of
matching is now replaced by a single comparison of
the classification result with the GIS database without
using any thresholds.
1.1. Related work
This kind of approach is an object-oriented image
analysis that is also successfully applied to other
Fig. 2. Differences between object-based and pixel-based classification.
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238 227
problems. A good overview of different approaches
can be found in Blaschke et al. (2000). These
approaches can be subdivided into approaches that
use existing GIS data to superimpose it on an image
(per-field or per-parcel classification), and approaches
that use object-oriented classification rules without
any GIS input. Approaches that use existing GIS data
are not very widely used today. In Aplin et al. (1999),
an example for a per-field classification approach is
introduced, which first classifies the image into
different land-use classes. Afterwards, the fields
(which represent forest parcels from a GIS database)
are subdivided into different classes, depending on
the classification result, by using thresholds. The
main difference of existing approaches compared
with our approach is that no thresholds are used in
our approach.
2. Object-based classification
2.1. Input data
The following tests were carried out with ATKIS
datasets. ATKIS is the German national topographic
and cartographic database, and captures the landscape
in the scale of 1:25,000 (AdV, 1988). In Walter
(1999), it was shown that a spatial resolution of at
least 2 m is needed to update data in the scale of
1:25,000. The remote sensing data were captured with
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238228
the DPA system, which is an optical airborne digital
camera (Hahn et al., 1996). The original resolution of
0.5 m was resampled to a resolution of 2 m. The DPA
system has four multispectral channels [blue 440–525
nm, green 520–600 nm, red 610–685 nm, near-
infrared (NIR) 770–890 nm].
2.2. Classification classes
Currently, 63 different object classes are collected
in ATKIS. There are a lot of object classes that can
have very similar appearances in an image of 2 m
pixel size (e.g., industrial areas, residential areas, or
areas of mixed use). Therefore, we do not use 63 land-
use classes for the classification but subdivide all
object classes into the five land-use classes: water,
forest, settlement, greenland, and roads. The land-use
Fig. 3. Input data for (a) object-based a
class roads is only used in the first step in the process
for the pixel-based classification. Because of the
linear shape, roads consist of many mixed pixels in
a resolution of 2 m and have to be checked with other
techniques (see Walter, 1998).
2.3. Input channels
Like in a pixel-based classification, we can use all
spectral bands as input channels. The difference is that
in the pixel-based classification, each pixel is classi-
fied separately, whereas in the object-based classifi-
cation, all pixels that belong to one GIS object are
grouped together. In order to analyse the spectral
behaviour of objects, we calculate the mean grey
value of each channel for all GIS objects. Fig. 3
shows as an example the original input data (b) and
nd (b) pixel-based classification.
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238 229
the mean RGB (red green blue) value (a) of each GIS
object. The result of the pixel grouping is like a
smoothing of the data. The spectral behaviour of the
objects is similar to the typical spectral behaviour of
the pixels. For example, forest areas are represented in
the green channel by dark pixel/objects, whereas
settlements are represented by bright pixel/objects.
This behaviour can be also seen in Fig. 4. The
scatterplots show the distribution of (a) the grey values
of settlement and forest pixels compared with the
distribution of (b) the mean grey value of settlement
and forest objects in the channels red and NIR. It can
be seen that the behaviour is similar but the separation
of the two classes becomes blurred because of the
smoothing effect. In the object-based classification, all
multispectral bands of the DPA camera system (blue,
green, red, and NIR) are used as input channels.
Fig. 4. Scatterplot of (a) p
Different land-use classes cannot be distinguished
only by their spectral behaviour but also by their
different textures. Texture operators transform input
images in such a way that the texture is coded in grey
values. In our approach, we use a texture operator
based on a co-occurrence matrix that measures the
contrast in a 5� 5 pixel window. Fig. 5 shows the
used texture operator in an example. The input image
is shown in Fig. 5a, the texture (calculated from the
blue band) in Fig. 5b, and the average object textures
in Fig. 5c. Settlements are represented with dark
pixels, greenlands with bright pixels, and forests with
middle grey pixels.
The variance of the grey values of the pixels of an
object is also a good indicator of the roughness of a
texture. Fig. 6 shows the calculated mean variance in
the blue band for all objects. Settlement objects have
ixels vs. (b) objects.
Fig. 5. (a) Input image, (b) texture blue band, and (c) average object texture.
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238230
high variance, greenland objects have middle variance,
and forest objects have low variance. Fig. 7 shows the
behaviour of the variance in the different bands: blue,
green, red, and NIR. The best discrimination between
land-use classes using the variance can be seen in the
blue band. In the NIR band, all land-use classes have a
similar distribution, which makes discrimination in this
band impossible.
Vegetation indices are very often used in pixel-
based classification as an input channel to improve
the classification result. They are based on the spectral
behaviour of chlorophyll, which absorbs red light and
reflects NIR light. In our approach, we employ the most
widely used normalised difference (Campbell, 1987):
VI ¼ IR� R
IRþ Rð1Þ
Fig. 8a shows the calculated vegetation index for
pixels and Fig. 8b for objects. It can be seen that
Fig. 6. Mean variance of GIS objects in blue band.
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238 231
settlements are represented typically by dark areas,
whereas forests are represented mostly by bright
areas. The classification of greenlands is difficult
because they can be represented by very bright areas
(e.g., fields with a high amount of vegetation) as well
as by very dark areas (e.g., fields shortly after the
harvest).
All so far defined input channels are also used in
‘‘normal’’ pixel-based classification. In object-based
classification, it is possible to add further input
channels, which do not describe directly spectral or
textural characteristics. For example, we can use the
result of a pixel-based classification and count the
percentage of pixels that are classified to a specific
land-use class. This evaluation is shown in Fig. 9. The
input image is shown in Fig. 9a and the pixel-based
classification result in Fig. 9b. Fig. 9c shows for each
object the percentage of pixels that are classified to
the land-use class forest. White colour represents
100% and black colour represents 0%. In Fig. 9b
and c, it can be seen that forest is a land-use class that
can be classified with high accuracy in pixel-based as
well as object-based classifications. Fig. 9d shows the
percentage of settlement pixels. Because of the high
resolution (2 m) of the data, settlements cannot be
detected as homogenous areas but they are split into
different land-use classes depending on what the
pixels are actually representing. Therefore, settlement
objects contain typically only 50–70% settlement
pixels in 2-m resolution images. This can be also seen
in Fig. 9e, which shows the percentage of greenland
pixels. Whereas greenlands contain up to 100% green-
land pixels, it can be seen that, in settlement areas,
pixels are also classified as greenlands.
An interesting visualisation of the feature space of
the object-based classification can be made with the
combination of three object-based evaluations of the
pixel-based classification. In Fig. 10, the percentage
of settlement pixels is assigned to the red band, the
percentage of forest pixels to the green band, and the
percentage of greenland pixels to the blue band of an
RGB image. The combination of these three bands
shows that the pixel-based classification of forests and
greenlands is very reliable, which can be seen on the
bright green and blue colour of the corresponding
objects. Settlement areas in contrast cannot be classi-
fied as homogenous areas. Therefore, settlement
objects are represented in a reddish colour that can
be brownish or purple.
3. Classification results
The approach was tested on two test areas (16 and
9.1 km2), which were acquired at different dates with
a total of 951 objects (194 forests, 252 greenlands,
497 settlements, and 8 water objects). The input
channels were:
� mean grey value blue band� mean grey value green band� mean grey value red band� mean grey value NIR band� mean grey value vegetation index� mean grey value texture from blue band� variance blue band� variance green band� variance red band
Fig. 7. Object variance in different bands (x-axis, variance; y-axis,
number of objects).
V. Walter / ISPRS Journal of Photogrammetry & Remote Sensing 58 (2004) 225–238232