Towards soft classification of satellite data A case study based ... - KTH
Towards soft classification of satellite data A case study based ... - KTH
Towards soft classification of satellite data A case study based ... - KTH
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
<strong>Towards</strong> <strong>s<strong>of</strong>t</strong> <strong>classification</strong> <strong>of</strong> <strong>satellite</strong> <strong>data</strong><br />
A <strong>case</strong> <strong>study</strong> <strong>based</strong> upon Resurs MSU-SK <strong>satellite</strong> <strong>data</strong> and land<br />
cover <strong>classification</strong> within the Baltic Sea Region.<br />
Anna Haglund<br />
Master <strong>of</strong> Science Project no. In Geoinformatics<br />
Royal Institute <strong>of</strong> Technology<br />
Department <strong>of</strong> Geodesy and Photogrammetry<br />
Stockholm, Sweden<br />
March 2000
Universitetsservice US AB<br />
Stockholm 2000<br />
08-790 7400
Preface<br />
This master <strong>of</strong> science project has been performed at the department <strong>of</strong> geodesy and<br />
photogrammetry, division for geoinformatics by assignment <strong>of</strong> Satellus, during the autumn<br />
1999. The project comprises 20 credits and is the completion <strong>of</strong> the education for technical<br />
surveyors at the Surveying Program at the Royal Institute <strong>of</strong> Technology, <strong>KTH</strong>, in Stockholm.<br />
I would like to thank my supervisors at <strong>KTH</strong>: Sindre Langaas, dept.<strong>of</strong> Civil and<br />
Environmental Engineering and Maria Roslund, dept. <strong>of</strong> Geodesy and Photogrammetry.<br />
I would also like to thank Satellus in Solna, where I have performed the project.<br />
Especially I want to thank Kjell Wester, who has been my supervisor at Satellus and has<br />
given many technical advises and come up with ideas.<br />
I<br />
Stockholm, mars 2000<br />
Anna Haglund
Sammanfattning<br />
Vid användning av medelupplösande satellitbilder med en pixelstorlek på 150-250 m på<br />
marken kan det vara svårt att välja homogena träningsytor. Pixelstorleken är också större än<br />
många objekt på marken, varför de kanske försvinner i en ”hård klassning”, där varje pixel<br />
bara kan tilldelas en klass. Ibland vill man att varje pixel ska innehålla så många klasser som<br />
det är i verkligheten. För att klara dessa problem har s k. mjuka klassificerare utvecklats.<br />
Syftet med rapporten är att undersöka hur många klasser som kan fås från<br />
medelupplösande satellit<strong>data</strong> när man använder högupplösande <strong>data</strong> som referens för<br />
generering av träningsytor och fastställa hur bra klassificeringen blir.<br />
Undersökningen gjordes med delar av två RESURS-bilder från två olika tidpunkter, med<br />
tre spektrala band, över en del av södra Sverige. Som referens användes svenska<br />
landtäcke<strong>data</strong> (SLD). För fyra olika samplingsmetoder jämfördes resultatet från IDRISIs<br />
moduler för mjuka klassificerare med varandra. Det skattade medelfelet varierade mellan<br />
klasserna och var minst för de små klasserna och förekomsten för varje klass jämfört med<br />
referensbilden skiljde sig relativt mycket. En orsak till den bristfälliga överensstämmelsen kan<br />
delvis bero på att referensbilden inte är helt korrekt. Referensbilden är baserad på information<br />
från den digitala topografiska kartan, varifrån klassgränser tagits och klassning <strong>of</strong>tast skett<br />
bara innanför gränsen med maximum-likelihood. Med antagandet att referensbilden är korrekt<br />
kommer felaktigheten i referens<strong>data</strong> att skyllas på RESURS-klassningen och därmed försämra<br />
den bedömda noggrannheten. Passningen mellan RESURS-scenerna och referensbilden är<br />
också en felkälla liksom passningen RESURS-scenerna sinsemellan. Så det finns både<br />
tematiska och geometriska fel. En annan orsak är för få träningsytor tillsammans med att<br />
klasserna inte alltid är spektralt separerbara.<br />
En förbättring av mjukvaran måste också göras innan någon användning i produktion kan<br />
bli aktuell.<br />
III
Abstract<br />
When using medium resolution <strong>satellite</strong> images with a pixel size <strong>of</strong> 150-250 m on the ground<br />
it can be difficult to collect homogenous training sites. The pixel size is also larger than many<br />
objects on the ground, why they might disappear in a “hard <strong>classification</strong>”, where every pixel<br />
just can contain one class. The wish is <strong>of</strong>ten that every pixel should contain as many classes<br />
as there are in reality. To manage these problems, so called <strong>s<strong>of</strong>t</strong> classifiers have been<br />
developed.<br />
The purpose <strong>of</strong> the paper is to examine how many classes that can be extracted from<br />
medium resolution <strong>satellite</strong> <strong>data</strong> when using higher resolution <strong>satellite</strong> <strong>data</strong> as reference and<br />
how well the <strong>classification</strong> performs.<br />
This examination was conducted using parts <strong>of</strong> two RESURS images from two different<br />
dates, with three spectral bands, covering a section <strong>of</strong> the southern parts <strong>of</strong> Sweden. As<br />
reference the national land cover <strong>classification</strong> was used. For four different samplings I<br />
compared the results from the IDRISI modules for <strong>s<strong>of</strong>t</strong> <strong>classification</strong> with each other. The<br />
RMS errors varied between the classes and were smallest for the small classes and the<br />
occurrence in the image for each class compared to the reference image differed relatively<br />
much. One reason for the insufficient agreement can partly be due to that the reference image<br />
is not completely correct. The reference image is <strong>based</strong> on information from the digital<br />
topographical map, from which class boundaries has been taken and the <strong>classification</strong> mostly<br />
been performed within these, with maximum-likelihood. With the assumption that the<br />
reference image is correct, the errors in this will be blamed on the RESURS <strong>classification</strong> and<br />
thus worsen the assessed accuracy. The geometric correction between the RESURS images<br />
and the reference image, as well as the geometric correction between the RESURS images<br />
them selves, are also an error source. So there are both thematic and geometric errors. One<br />
other reason is too few training sites along with not always spectrally separable classes.<br />
There need to be done some improvements in the <strong>s<strong>of</strong>t</strong>ware before any use in production can<br />
be <strong>of</strong> interest.<br />
IV
Contents<br />
PREFACE...................................................................................................................................................................I<br />
SAMMANFATTNING.......................................................................................................................................... III<br />
ABSTRACT.............................................................................................................................................................IV<br />
CONTENTS ............................................................................................................................................................. V<br />
1. INTRODUCTION ................................................................................................................................................1<br />
1.1BACKGROUND...................................................................................................................................................1<br />
1.2 OBJECTIVES......................................................................................................................................................1<br />
1.3 THE BALANS PROJECT ..................................................................................................................................1<br />
1.4 STRUCTURE OF THE THESIS..............................................................................................................................2<br />
2. MEDIUM RESOLUTION SENSORS...............................................................................................................3<br />
2.1 AVAILABLE SENSORS .......................................................................................................................................3<br />
2.2 FUTURE SENSORS .............................................................................................................................................4<br />
3. A REVIEW OF SOFT CLASSIFIERS..............................................................................................................5<br />
3.1 INTRODUCTION.................................................................................................................................................5<br />
3.2 FUZZY SET THEORY.........................................................................................................................................6<br />
3.2.1 Fuzzy c-means (k-means) ........................................................................................................................7<br />
3.2.2 Applications with Fuzzy c-means classifier............................................................................................7<br />
3.3 SOFTENING MAXIMUM LIKELIHOOD CLASSIFIER (MCL).................................................................................8<br />
3.3.1 Applications with <strong>s<strong>of</strong>t</strong>ening MCL ...........................................................................................................8<br />
3.4 LINEAR MIXTURE MODELLING.........................................................................................................................8<br />
3.4.1 Applications with linear mixture modelling ...........................................................................................9<br />
3.5 METHODS BASED ON BAYESIAN PROBABILITY THEORY .................................................................................9<br />
3.5.1 Applications with Bayesian probability theory ....................................................................................10<br />
3.6 METHODS BASED ON DEMPSTER-SHAFER THEORY ......................................................................................10<br />
3.6.1 Applications with Dempster-Shafer theory ..........................................................................................11<br />
3.7 NEURAL NETWORK........................................................................................................................................11<br />
3.7.1 The Multilayer Perceptron Classifier using feed-forward and back-propagation.............................11<br />
3.7.2 Applications with Multilayer Perceptron Classifier ............................................................................13<br />
3.7.2 Neural Network using Cascade-correlation.........................................................................................14<br />
3.7.3 Applications with Cascade-correlation................................................................................................14<br />
3.7.4 Fuzzy ARTMAP......................................................................................................................................14<br />
3.7.5 Applications with Fuzzy ARTMAP........................................................................................................16<br />
4. DATA AND METHODS ...................................................................................................................................17<br />
4.1 INTRODUCTION...............................................................................................................................................17<br />
4.2 DATA ..............................................................................................................................................................17<br />
4.2.1 Satellite <strong>data</strong> ..........................................................................................................................................17<br />
4.2.2 Reference <strong>data</strong> .......................................................................................................................................18<br />
4.3 STUDY AREA...................................................................................................................................................19<br />
4.3.1 General description <strong>of</strong> the Baltic Sea region landscape......................................................................19<br />
4.3.2 Quantitative characterisation ...............................................................................................................20<br />
4.4 TOOLS.............................................................................................................................................................20<br />
4.4.1 IDRISI ....................................................................................................................................................20<br />
4.4.2 ERDAS IMAGINE .................................................................................................................................21<br />
4.5 METHODOLOGY..............................................................................................................................................22<br />
4.5.1 Outline....................................................................................................................................................22<br />
4.5.2 Initial <strong>data</strong> preprocessing .....................................................................................................................22<br />
4.5.3 Choice <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers .......................................................................................................................24<br />
V
4.5.4 Application <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers................................................................................................................24<br />
4.5.4 Evaluation and comparison <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers .....................................................................................25<br />
5. RESULTS ............................................................................................................................................................27<br />
5.1 INTRODUCTION...............................................................................................................................................27<br />
5.2 FUZZY SIGNATURES .......................................................................................................................................27<br />
5.3 RMS ERRORS .................................................................................................................................................27<br />
5.4 OVER- AND UNDER CLASSIFIED PIXELS .........................................................................................................29<br />
5.5 AVERAGE VALUES..........................................................................................................................................33<br />
6. DISCUSSION......................................................................................................................................................40<br />
7. CONCLUSIONS AND RECOMMENDATIONS..........................................................................................41<br />
REFERENCES.........................................................................................................................................................42<br />
APPENDIX 1: ACTION STEPS FOR SOFT CLASSIFICATION...............................................................................44<br />
APPENDIX 2: MACROFILE- SCHEME ..............................................................................................................47<br />
VI
1. Introduction<br />
1.1 Background<br />
Satellite images have been used for land cover <strong>classification</strong> ever since the first earth resource<br />
<strong>satellite</strong>s were launched. The most used method for <strong>classification</strong> has been the maximumlikelihood<br />
algorithm. Recently, medium resolution <strong>satellite</strong> <strong>data</strong> (150-250 m) has begun to be<br />
provided. Also then was the maximum-likelihood algorithm used in <strong>classification</strong>. There have<br />
been several studies done with the maximum-likelihood classifier. One made with medium<br />
resolution <strong>satellite</strong> <strong>data</strong> was done by Malmberg and Furberg (1997). The last five years there<br />
has been a progress on the method side. New classifiers have been developed – <strong>s<strong>of</strong>t</strong><br />
classifiers. They have been developed as a consequence that coarse- and medium resolution<br />
<strong>satellite</strong> <strong>data</strong> <strong>of</strong>ten contain mixed pixels, e.g. forests have no sharp boundaries between<br />
coniferous- and deciduous forest.<br />
It is a new and exciting field. These new <strong>s<strong>of</strong>t</strong> classifiers can extract more information than<br />
the maximum-likelihood classifier.<br />
1.2 Objectives<br />
The main objective is to evaluate and compare the suitability <strong>of</strong> some <strong>s<strong>of</strong>t</strong> classifiers to<br />
deduce land cover information in a patchy landscape using medium resolution <strong>satellite</strong> <strong>data</strong>,<br />
in particular for a <strong>study</strong> area in the Baltic Sea region for the purpose <strong>of</strong> the BALANS project<br />
(Landcover Information for the Baltic Sea Drainage Basin).<br />
Secondary objectives are to provide information about present and future <strong>satellite</strong>s with<br />
sensors that provide medium resolution <strong>data</strong> and make a review <strong>of</strong> the available <strong>s<strong>of</strong>t</strong><br />
classifiers today and there requirements on input- and reference <strong>data</strong>.<br />
1.3 The BALANS Project<br />
The Remote Sensing Technology Division at Satellus, a subsidiary to the Swedish Space<br />
Corporation, has in partnership with six organisations in Sweden, Norway, Finland and<br />
Poland, started a project called BALANS. The BALANS project’s aim is to develop a<br />
<strong>satellite</strong>-<strong>based</strong> land cover <strong>data</strong>base for the complete Baltic Sea Drainage Basin, see figure 1,<br />
using medium resolution <strong>satellite</strong>s (Olsson 1999). Many organisations in the Baltic Sea region<br />
have a common need for basic geographical <strong>data</strong> sets. Up to now the main need has been from<br />
the environmental, hydrometeorological and marine sciences, but now there are growing<br />
requirements from the physical planning, political and business communities (Langaas 1996).<br />
Today, there are many growing pressures on the Baltic Sea. There is a critical need for<br />
development that is sustainable, both economically and environmentally.<br />
The <strong>data</strong>base is envisaged to have a maximum resolution in the order <strong>of</strong> 200 meters.<br />
The project will last for three years, and will be finished sometime in the end <strong>of</strong> year 2001<br />
or the beginning <strong>of</strong> year 2002.<br />
1
Figure 1. The Baltic Drainage Basin and the BALANS <strong>study</strong> area<br />
(adapted from http://www.grida.no/baltic/htmls/maps.htm).<br />
1.4 Structure <strong>of</strong> the thesis<br />
In chapter 2, a review <strong>of</strong> present and future <strong>satellite</strong>s with medium resolution <strong>satellite</strong> <strong>data</strong> is<br />
provided. Chapter 3 is a review <strong>of</strong> available <strong>s<strong>of</strong>t</strong> classifiers and some applications made with<br />
those. Chapter 4 contains the <strong>data</strong> and method used in this <strong>study</strong>, chapter 5 the result and<br />
discussion, followed by chapter 6 with the conclusions.<br />
2
2. Medium Resolution Sensors<br />
2.1 Available sensors<br />
Medium resolution sensors are sensors that have a pixel size covering the ground <strong>of</strong> about<br />
150-250 meter. They bridge the gap in image extent and spatial resolution between<br />
SPOT/Landsat TM (20-30 m) and NOAA AVHRR (1.1 km) and are appropriate for mapping<br />
<strong>of</strong> large areas, since not so many images are needed and the images are not too coarse.<br />
Two <strong>satellite</strong>s with medium resolution sensors covering the Nordic area are the Russian<br />
<strong>satellite</strong> RESURS-O1 and the Indian Remote Sensing <strong>satellite</strong>, IRS. The <strong>satellite</strong>s have<br />
different sensors. RESURS-O1 has a MSU-SK sensor and IRS has a wide field sensor (WIFS),<br />
see descriptions and <strong>data</strong> in table 1. The only available MSU-SK <strong>data</strong> today is archived<br />
RESURS-O1 #3 <strong>data</strong> (SSC Satellitbild 1999). One orbit period takes 98 minutes and the<br />
repetition cycle for that <strong>satellite</strong> is 21 days, with a potential coverage <strong>of</strong> a specific area every<br />
third to fourth day at the equator. Every scene covers 600 x 600 km 2 . It is a wide-angle<br />
instrument with a conical scan where every pixel has the same size and viewing angle. The<br />
<strong>data</strong> from the curved scanning lines are resampled by cubic convolution from the nominal 170<br />
meters raw pixels to 160 meter pixels to obtain a quadratic image grid.<br />
The WIFS has two spectral bands, the red and the near infrared, with the applications to see<br />
chlorophyll absorption <strong>of</strong> plants and to survey biomass and see waterbodies, respectively<br />
(euromap 1997). One orbit around the earth takes 101.35 minutes to complete and it takes 341<br />
orbits to cover the entire earth. The repetition cycle for the IRS <strong>satellite</strong>s is 24 days, but IRS<br />
1C/D works as a system so their combined repetition is 12 days. At the Nordic latitudes a<br />
specific area is covered every 1-2 days. Also the WIFS’s scanning lines are resampled by cubic<br />
convolution to obtain a quadratic image grid.<br />
Both <strong>satellite</strong> types have large swath width. The largest difference between the <strong>satellite</strong>s is<br />
the number <strong>of</strong> wavelength bands. While RESURS-O1 MSU-SK has five bands: red, green, two<br />
near infrared (NIR) and one thermal infrared (TIR), IRS WIFS only has two: the red and near<br />
infrared. But in practice the TIR band in MSU-SK cannot be acquired in parallel with the<br />
others. Experiences from the <strong>data</strong> indicates that there are better radiometric dynamics in <strong>data</strong><br />
from WIFS on IRS than in the <strong>data</strong> from MSU-SK on RESURS-O1 #3 (Hyyppä 1999).<br />
3
Table 1. Summary <strong>of</strong> <strong>satellite</strong> and sensor <strong>data</strong>.<br />
Satellites:<br />
RESURS-O1 MSU-SK IRS-1C WiFS IRS-1D WiFS<br />
Orbit sun-synchronous, circular sun-synchronous, circular, near polar<br />
Average altitude 678 km 817 km<br />
Inclination 98.04 deg 98.69 deg<br />
Orbit period 98 min 101.35 min<br />
Orbit repeat cycle 21 days 12 days (24 days for each)<br />
Launch date 4/11 1994 28/12 1995 29/12 1997<br />
Sensors:<br />
Imaging mechanism Conical scan 2048 element linear array CCD<br />
Viewing angle 39 deg ?<br />
Spectralband: Wavelength: pixelsize: Wavelength: pixelsize:<br />
1: Green 0.5-0.6 µm 160 m -<br />
2: Red 0.6-0.7 µm 160 m 0.62-0.68 µm 188 m<br />
3:NIR 0.7-0.8 µm 160 m 0.77-0.86 µm 188 m<br />
4: NIR 0.8-1.1 µm 160 m -<br />
5: TIR 10.4-12.6 µm 600 m -<br />
Swath width 600 km 810 km 728-812 km<br />
Potential coverage <strong>of</strong> same area 3-4 days at the equator 1-2 days at Nordic latitudes<br />
Overlap ? Ca. 80%<br />
2.2 Future sensors<br />
There are two new medium resolution sensors coming. These are MODIS on the EOS AM-1<br />
and MERIS on the ENVISAT <strong>satellite</strong> (Hyyppä 1999).<br />
MODIS has a cross-track scan mirror and a set <strong>of</strong> linear arrays with spectral interference<br />
filters located in four focal planes. It has a viewing swath width <strong>of</strong> 2330 km and will cover the<br />
earth every 1-2 days. It will have 36 spectral bands in the range <strong>of</strong> 0.4-14.4 µm optimized for<br />
measuring surface temperature, ocean colour, global vegetation, cloud characteristics, aerosol<br />
concentrations, temperature and moisture soundings, snow cover and ocean currents. The<br />
spatial resolution will be 250 m for band 1-2, 500 m for band 3-7 and 1000 m for band 8-36.<br />
EOS AM-1 MODIS <strong>data</strong> will be available in spring 2000.<br />
MERIS has radiation-sensitive arrays (CCD’s) that provides spatial sampling in the acrosstrack<br />
direction and the <strong>satellite</strong>’s motion provides scanning in the along- track direction. It<br />
will have 15 spectral bands in the range <strong>of</strong> 0.39-1.04 µm, that can be selected by ground<br />
command. The swath width will be 1150 km. The spatial resolution will be 300 m over<br />
coastal zones and land surfaces where communication capabilities exists. Otherwise, the<br />
resolution is reduced to 1200 m to reduce the amount <strong>of</strong> <strong>data</strong> that will be recorded onboard.<br />
ENVISAT MERIS <strong>data</strong> will be available in spring 2001.<br />
4
3. A review <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />
3.1 Introduction<br />
In conventional <strong>classification</strong> in remote sensing discrete pixels are used, i.e. the result is only<br />
one class per pixel. Much information about the memberships <strong>of</strong> the pixel to other classes is<br />
lost. This additional information can increase the accuracy <strong>of</strong> the <strong>classification</strong>.<br />
Mixed pixels or ‘mixels’ occur because the pixel size may not be fine enough to capture<br />
detail on the ground necessary for specific applications or where the ground properties, such<br />
as vegetation and soil types, vary continuously, as almost everywhere.<br />
Fuzziness suggests that a given pixel, owing to its spectral reflectance properties, may be<br />
placed into more than one informational/spectral class. The output <strong>of</strong> a <strong>s<strong>of</strong>t</strong> <strong>classification</strong> is a<br />
set <strong>of</strong> images (one per class) that express for each pixel the degree <strong>of</strong> membership in the class<br />
in question.<br />
S<strong>of</strong>t classifiers can be useful in delineating forest boundaries, shorelines and other<br />
continuous classes. They can also bring out objects that cover small areas, which with<br />
conventional classifiers otherwise would have disappeared.<br />
In training and testing in a <strong>classification</strong>, mixed pixels are usually avoided. But it may be<br />
difficult to acquire a training set <strong>of</strong> an appropriate size if only pure pixels are selected for<br />
training, since large homogenous regions <strong>of</strong> each class are needed in the image. The training<br />
statistics defined, may not be fully representative <strong>of</strong> the classes and so provide a poor base for<br />
the remainder <strong>of</strong> the analysis.<br />
By recognising that there are various degrees to which fuzziness may be incorporated in<br />
each stage, a continuum <strong>of</strong> <strong>classification</strong> fuzziness may be defined (Foody 1999). As<br />
fuzziness may be a characteristic feature <strong>of</strong> both the remotely sensed <strong>data</strong> and the ground<br />
<strong>data</strong>, the use <strong>of</strong> a fuzzy <strong>classification</strong> algorithm alone may be insufficient for the resolution <strong>of</strong><br />
the mixed-pixel problem.<br />
The continuum ranges from fully fuzzy, where fuzziness is accommodated in all three<br />
stages <strong>of</strong> the <strong>classification</strong> (the training-, allocation and testing stages), to completely crisp,<br />
which are the conventional <strong>classification</strong>.<br />
A modified maximum-likelihood classifier can accommodate fuzziness in any or all three<br />
stages <strong>of</strong> the analysis and neural networks can provide a <strong>classification</strong> at any point along the<br />
continuum <strong>of</strong> <strong>classification</strong> fuzziness.<br />
In training, the <strong>data</strong> is hard if the pixels are pure and vary in fuzziness for mixed pixels.<br />
Data on the fuzzy membership properties <strong>of</strong> training samples may be used to derive refined<br />
class descriptors for use in the class allocation stage. By refining the training statistics where<br />
some or even all <strong>of</strong> the pixels were mixed, the statistics that would have been derived if the<br />
pixels had in fact been pure can be obtained (Foody 1999).<br />
In the allocation stage for a <strong>s<strong>of</strong>t</strong> classifier a pixel can be allocated from a single class to all<br />
classes it provide membership in.<br />
This chapter will describe the theory and algorithms behind the existing <strong>s<strong>of</strong>t</strong> classifiers.<br />
To make the theory more understandable this chapter will also describe some applications that<br />
have been made.<br />
5
3.2 Fuzzy Set Theory<br />
Fuzzy sets are sets without sharp boundaries and they are applied to handle uncertainty in the<br />
process <strong>of</strong> <strong>classification</strong> (Palubinskas 1994). The result is <strong>of</strong>ten a more detailed and precise<br />
<strong>classification</strong>. Fuzziness can effectively extend the usefulness <strong>of</strong> map products developed<br />
from remote sensing imagery. The fuzzy set theory is particularly interesting as the analyst<br />
controls the degree <strong>of</strong> fuzziness (Foody 1996).<br />
It is the shape <strong>of</strong> the membership function that defines the ‘fuzziness’ <strong>of</strong> the phenomenon<br />
represented by the set. There are four membership functions that are usually used: sigmoidal<br />
(s-shaped), J-shaped, linear and user-defined (Eastman 1997). The sigmoidal membership<br />
function is the most commonly used in fuzzy set theory, see figure 2. It can be produced using<br />
a cosine function that requires the positions (along the X-axis) <strong>of</strong> 4 points governing the shape<br />
<strong>of</strong> the curve.<br />
The J-Shaped membership function is also quite common, although in most <strong>case</strong>s the<br />
sigmoidal function would be better. The linear is used extensively in electronic devices<br />
advertising fuzzy set logic and the user-defined is used when the relationships between the<br />
value and fuzzy membership does not follow any <strong>of</strong> the above three functions. The userdefined<br />
is the most applicable (Eastman 1997). The control points used in this function can be<br />
as many as necessary to define the fuzzy membership curve. The fuzzy membership between<br />
any two control points is linearly interpolated.<br />
Figure 2. The Sigmoidal Membership Function (<strong>based</strong> on Eastman 1997).<br />
The outputs for each pixel are the different membership values for each class. A high<br />
membership value indicates that the relevant land cover type covers a large area <strong>of</strong> the pixel<br />
(Bastin 1997). When using summation <strong>of</strong> memberships for a post-<strong>classification</strong> <strong>of</strong> fuzzy<br />
<strong>classification</strong> results, most <strong>of</strong> the increase <strong>of</strong> average <strong>classification</strong> accuracy is achieved after<br />
the first iteration (Palubinskas 1994).<br />
6
3.2.1 Fuzzy c-means (k-means)<br />
Fuzzy c-means is a clustering algorithm used for either unsupervised or supervised<br />
<strong>classification</strong>. The algorithm subdivides a <strong>data</strong> set into c-clusters or classes (Foody 1996). It<br />
begins by randomly assigning pixels to classes and iteratively moves the pixels to other<br />
classes with the aim <strong>of</strong> minimizing the generalised least-squared error,<br />
J m ( U,<br />
v)<br />
= ∑∑(<br />
u<br />
n<br />
c<br />
k = 1 i=<br />
1<br />
where U is a fuzzy c-partition <strong>of</strong> the <strong>data</strong> set Y containing n pixels, c is the number <strong>of</strong> classes,<br />
|| ||A is an inner product norm, v is a vector <strong>of</strong> cluster centres, vi is the centre <strong>of</strong> cluster i and<br />
m is a weighting component that lies within the range 1≤ m ≤ ∞ which determines the degree<br />
<strong>of</strong> fuzziness. The inner product norm is derived from,<br />
2<br />
7<br />
k<br />
)<br />
m<br />
y<br />
k<br />
− v<br />
T<br />
y − v = ( y − v ) A(<br />
y − v )<br />
k<br />
i<br />
A<br />
k<br />
−1<br />
A number <strong>of</strong> norms may be selected, for instance the Mahalanobis norm, A = C y , where<br />
Cy is the covariance matrix <strong>of</strong> the <strong>data</strong> set Y.<br />
To implement the fuzzy c-means algorithm, additional parameters are required to guide the<br />
partitioning process. These parameters are: selection <strong>of</strong> a distance measure and choosing a<br />
weighting exponent. The weighting exponent controls the ‘hardness’ or ‘fuzziness’ <strong>of</strong> the<br />
<strong>classification</strong>. The range <strong>of</strong> useful values for m is 1.5 < m < 3.0.<br />
The fuzzy c-means algorithm is particularly useful in circumstances where it is not<br />
reasonable to make assumptions about the statistical distributions <strong>of</strong> sample <strong>data</strong> (e.g. where<br />
training sets <strong>of</strong> pure pixels are small).<br />
For each pixel a fractional value is obtained for each class in the form <strong>of</strong> a real number<br />
between 0 and 1, and will generally sum up to 1.0 across all candidate classes.<br />
As output one also gets a residual error map showing the RMS error across the image.<br />
The values need to be adjusted (i.e. using the slope from linear regression) for area<br />
predictions to be made.<br />
3.2.2 Applications with Fuzzy c-means classifier<br />
A comparative <strong>study</strong> between traditional unsupervised <strong>classification</strong> and the fuzzy c-means<br />
algorithm was done by Manyara and Lein (1994). Landsat MSS subsets were conducted to<br />
detect environmental change. The fuzzy c-means were used since the boundary between forest<br />
and other categories <strong>of</strong> land cover may be difficult to delineate precisely. Four smaller <strong>study</strong><br />
sites depicting different forest environments were extracted from the image. The approach <strong>of</strong><br />
the fuzzy c-means proved that a pixel <strong>of</strong>ten belongs to more than one class. Of the total 2 500<br />
pixels per site, only 249 were pure in site I, 219 in site II, 29 in site III and 87 in site IV.<br />
Another comparison has been made by Bastin (1997) between the fuzzy c-means and the<br />
linear mixture model in the use for unmixing <strong>of</strong> coarse pixel signatures to identify four land<br />
cover classes. The Mahalanobis norm was used as a typically measure and after a little<br />
experimentation the weighting exponent, m, was set to 1.5. By this, dominant and minor land<br />
cover classes in each pixel was easily identified, without the <strong>classification</strong> being overly hard.<br />
The fuzzy c-means classifier performed best overall at locating and quantifying inclusions in<br />
mixed pixels.<br />
i<br />
k<br />
i<br />
i<br />
2<br />
A
3.3 S<strong>of</strong>tening maximum likelihood classifier (MCL)<br />
MCL operate by using band means and standard deviations from training <strong>data</strong> to model land<br />
cover classes as centroids in feature space, surrounded by probability contours. The<br />
probability density function assumes that the sample values for each class are normally<br />
distributed. The unclassified pixels are plotted in the same feature space and get a posteriori<br />
probability. Usually the pixels are then assigned to the class for which they have highest<br />
membership probability. But it is possible to <strong>s<strong>of</strong>t</strong>en the maximum likelihood <strong>classification</strong> by<br />
using the a posteriori membership probability values as indices <strong>of</strong> class membership (Bastin<br />
1997).<br />
The <strong>data</strong> must still satisfy the assumptions and requirements <strong>of</strong> the <strong>classification</strong> technique<br />
used, which is <strong>of</strong>ten unlikely with the probability-<strong>based</strong> classifiers.<br />
3.3.1 Applications with <strong>s<strong>of</strong>t</strong>ening MCL<br />
In the comparison made by Bastin (1997) described above, a shell script was used to combine<br />
each set <strong>of</strong> eight maximum likelihood <strong>classification</strong> output maps, and to produce a map<br />
showing the a posteriori probabilities for each class. The result was mostly extremely<br />
classified pixels, i.e. most <strong>of</strong> them were classified as pure pixels or as habitats being entirely<br />
absent. To avoid an incorrect evaluation <strong>of</strong> this method, the <strong>classification</strong> should be <strong>s<strong>of</strong>t</strong>ened,<br />
by broadening the spread <strong>of</strong> the normal distribution functions, which represent the land cover<br />
classes.<br />
3.4 Linear mixture modelling<br />
For linear unmixing <strong>of</strong> the spectral <strong>data</strong>, a ‘pure’ spectral signature needs to be obtained from<br />
the training <strong>data</strong>, for each <strong>of</strong> the defined land cover types (Bastin 1997). The model views<br />
mixed pixel signatures as being made up <strong>of</strong> a simple weighted linear sum <strong>of</strong> the component<br />
land cover signatures in that pixel. The weights are directly determined by the relative<br />
proportions <strong>of</strong> ground covered by each different land cover class by,<br />
x = Mf + e<br />
where x is the set <strong>of</strong> digital values measured by the sensor in each <strong>of</strong> n measurement<br />
wavelengths, f is a vector <strong>of</strong> ground cover proportions, for each <strong>of</strong> c land covers.<br />
M is a matrix <strong>of</strong> n x c coefficients, in which the columns are the vectors µ1…µc, which<br />
represent the pure spectral signatures given by the c cover classes in the absence <strong>of</strong> noise, e is<br />
a terments noise, which is minimized in some way in order to obtain a measure <strong>of</strong> the<br />
proportion vector f.<br />
The expected signature for a mixed pixel is the sum,<br />
f1µ 1 + f 2µ<br />
2 + ... + f c µ c = Mf<br />
where for class c, fc is the proportion <strong>of</strong> the pixel covered, and µc is the characteristic<br />
signature.<br />
The noise caused by the sensor and by natural variability within a scene is quantified as<br />
the vector <strong>of</strong> errors, e, which can be quantified as the sum <strong>of</strong> squares <strong>of</strong> its elements or as the<br />
variance <strong>of</strong> the training <strong>data</strong>.<br />
The linear mixture model assumes that each photon on the ground is interacting with only<br />
one land cover type before being reflected back to the sensor. In a complex and non-random<br />
landscape the linear mixture model is a reasonable approximation.<br />
The model is trained using the matrix <strong>of</strong> values M generated from ground truth<br />
8
information on ground cover proportions within selected pixels or from selected ‘purest’<br />
pixels for each component. The effectiveness <strong>of</strong> the model is dependent on the degree <strong>of</strong><br />
separation <strong>of</strong> the different signatures and the level <strong>of</strong> noise present in the scene.<br />
Rather than incorporating the noise and variability <strong>of</strong> the sample <strong>data</strong> into statistical or<br />
fuzzy limits around each cluster centriod, the linear mixture model uses the centroid vectors<br />
to define the spectral signatures <strong>of</strong> the pure land cover classes, and separates out scene/sensor<br />
noise into a separate error term.<br />
The number <strong>of</strong> distinguishable land cover types is strictly limited by the number <strong>of</strong> spectral<br />
bands in the <strong>data</strong>.<br />
3.4.1 Applications with linear mixture modelling<br />
The comparison, mentioned above in the fuzzy c-means algorithm application and in the<br />
<strong>s<strong>of</strong>t</strong>ening <strong>of</strong> MCL, showed that the linear mixture model assigned extreme values to most <strong>of</strong><br />
the pixels (Bastin 1997). One <strong>of</strong> the reasons for that is that there were pixels outside the range<br />
from zero to one in the <strong>classification</strong>, so called outliers. These outliers were reclassified to<br />
minimum or maximum values respectively. The result from the linear mixture model was<br />
fairly successful at picking up the general spatial patterns and gradations <strong>of</strong> the different cover<br />
classes, but it had problems with specific land cover signatures, because <strong>of</strong> poor spectral<br />
separability (Bastin 1997).<br />
3.5 Methods <strong>based</strong> on Bayesian probability theory<br />
The prime use for methods <strong>based</strong> on Bayesian probability theory is to determine the extent to<br />
which mixed pixels exist in the image, and their relative proportions. The methods can only<br />
be used when complete information is available or assumed (Eastman 1997). They are <strong>based</strong><br />
on the bayesian probability theory, which is the primary tool for the evaluation <strong>of</strong> the<br />
relationship between the indirect evidence and the decision set. The bayesian probability<br />
theory allows us to combine new evidence about a hypothesis along with prior knowledge to<br />
arrive at an estimate <strong>of</strong> the likelihood that the hypothesis is true. The basis for this is<br />
Bayes´theorem, which states that,<br />
p(<br />
h e)<br />
=<br />
where:<br />
p(h⏐e) = the probability <strong>of</strong> the hypothesis being true, given the evidence (posterior<br />
probability)<br />
p(e⏐h) = the probability <strong>of</strong> finding that evidence, given the hypothesis, being true<br />
p(h) = the probability <strong>of</strong> the hypothesis being true regardless <strong>of</strong> the evidence (prior<br />
probability)<br />
∑<br />
i<br />
p(<br />
e h)<br />
⋅ p(<br />
h)<br />
p(<br />
e h ) ⋅ p(<br />
h )<br />
The variance/covariance matrix derived from training site <strong>data</strong> is the matrix that allows one<br />
to assess the multivariate conditional probability p(e⏐h). This quantity is then modified by the<br />
prior probability <strong>of</strong> the hypothesis being true and then normalized by the sum <strong>of</strong> such<br />
considerations over all classes. This modification- and normalization step is important<br />
because it assumes that the only possible interpretation <strong>of</strong> a pixel is one <strong>of</strong> those classes for<br />
which training site <strong>data</strong> have been provided. Thus even weak support for a specific<br />
interpretation may appear to be strong if it is the strongest <strong>of</strong> the possible choices given. So<br />
even if the reflectance <strong>data</strong> for a certain class in a pixel is very weak it is treated as<br />
unequivocally belonging to that class (p=1.0) if no support exists for other classes. It therefore<br />
admits no ignorance and is a confident classifier.<br />
9<br />
i<br />
i
The posterior probability p(h⏐e) is the same quantity that the maximum-likelihood module<br />
evaluates to determine the most likely class.<br />
If a pixel has posterior probabilities <strong>of</strong> belonging to certain classes <strong>of</strong> 0.72 and 0.28<br />
respectively, this would be interpreted as evidence that the pixel contains 72 % <strong>of</strong> the first<br />
class and 28 % <strong>of</strong> the second. But this requires that the classes for which training site <strong>data</strong><br />
have been provided are the only one existing and that the conditional probability distribution<br />
p(e⏐h) do not overlap in the <strong>case</strong> <strong>of</strong> pure pixels. These requirements may in practice be very<br />
difficult to meet.<br />
3.5.1 Applications with Bayesian probability theory<br />
No reports on studies using Bayesian probability theory have been found.<br />
3.6 Methods <strong>based</strong> on Dempster-Shafer Theory<br />
The methods <strong>based</strong> on Dempster-Shafer theory are the most important <strong>of</strong> the <strong>s<strong>of</strong>t</strong> classifiers.<br />
It allows for the expression <strong>of</strong> ignorance in uncertainty management (Eastman 1997). The<br />
prime use <strong>of</strong> these methods is to check for the quality <strong>of</strong> one’s training site <strong>data</strong> and the<br />
possible presence <strong>of</strong> unknown classes.<br />
The basic assumptions <strong>of</strong> Dempster-Shafer theory are that ignorance exists in the body <strong>of</strong><br />
knowledge, and that belief for a hypothesis is not necessarily the complement <strong>of</strong> belief <strong>of</strong> its<br />
negation.<br />
The degree to which evidence provides concrete support for a hypothesis is known as<br />
belief, and the degree to which the evidence does not refute that hypothesis is known as<br />
plausibility. The difference between these two is the belief interval, which acts as a measure<br />
<strong>of</strong> uncertainty about a specific hypothesis.<br />
If the evidence supports one class to the degree <strong>of</strong> 0.3 and all other to 0.0, the Dempster-<br />
Shafer theory would assign a belief <strong>of</strong> 0.3 to that class and a plausibility <strong>of</strong> 1.0, yielding a<br />
belief interval <strong>of</strong> 0.7. Furthermore, it would assign a belief <strong>of</strong> 0.0 to all other classes and a<br />
plausibility <strong>of</strong> 0.7.<br />
Dempster-Shafer theory defines hypotheses in a hierarchical structure and will accept all<br />
possible combinations <strong>of</strong> hypotheses, see figure 3. The combinations <strong>of</strong> hypotheses<br />
acceptance is because it <strong>of</strong>ten happens that the evidence we have, supports some<br />
combinations <strong>of</strong> hypotheses without the ability to further distinguish the subsets. The basic<br />
classes are called singletons and the combinations non-singletons.<br />
[A,B,C]<br />
[A,B] [A,C] [B,C]<br />
[A] [B] [C]<br />
Figure 3. The Hierarchical Structure <strong>of</strong> the Subsets in the Whole Set [A,B,C] (<strong>based</strong> on<br />
Eastman 1997).<br />
10
When evidence provides some degree <strong>of</strong> commitment to one <strong>of</strong> these non-singleton<br />
classes and not to any <strong>of</strong> its constituents separately, this commitment is called Basic<br />
Probability Assignment (BPA). Thus, belief in a non-singleton class is the sum <strong>of</strong> BPA’s for<br />
that class and all sub-classes. Belief represents the total commitment to all members <strong>of</strong> a set<br />
combined.<br />
These non-singletons represents mixtures and might be used for more detailed examination <strong>of</strong><br />
sub-pixel <strong>classification</strong>.<br />
3.6.1 Applications with Dempster-Shafer theory<br />
In a <strong>study</strong> <strong>of</strong> fire scars in Indonesia by Fuller and Fulky (1998), Dempster-Shafer theory was<br />
used. That algorithm was used because each AVHRR pixel had a resolution <strong>of</strong> 1.1 km and<br />
could have contained several patches <strong>of</strong> fire scars. The only training sites used was for fire<br />
scars, all other classes were to be ‘other’ in the <strong>classification</strong>. As input some surface<br />
temperature layers and vegetation change detection layers were used. No rigorous ground<br />
validation could be done, due to lack <strong>of</strong> ground <strong>data</strong> at a scale meaningful to the AVHRR<br />
analysis. But compared to the other method used in the <strong>study</strong>, multiple threshold approach,<br />
the Dempster-Shafer method managed to extract more fire scars and came closer to the result<br />
done by the the Center for Remote Sensing, Imaging and Processing (CRISP) <strong>of</strong> the National<br />
University <strong>of</strong> Singapore. The CRISP analysis relied on visual interpretation <strong>of</strong> 766 SPOT<br />
quicklook images mapped to 100 m spatial resolution.<br />
3.7 Neural Network<br />
Neural networks learn by example (Kanellopoulos 1997). From the training set the network<br />
can learn the values <strong>of</strong> the samples internal parameters. Usually neural networks deal with<br />
large amounts <strong>of</strong> training <strong>data</strong> (i.e. thousands <strong>of</strong> samples) whereas statistical methods use<br />
much smaller training sets. The goal is the representation <strong>of</strong> complicated phenomena. The<br />
neural network does not make any explicit a priori assumptions about the <strong>data</strong>.<br />
3.7.1 The Multilayer Perceptron Classifier using feed-forward and back-propagation<br />
The multilayer perceptron classifier (MLP) are the most widely used neural classifier today.<br />
These networks are general-purpose, flexible, non-linear models consisting <strong>of</strong> a number <strong>of</strong><br />
units organised into multiple layers (Kanellopoulos 1997). Varying <strong>of</strong> the number <strong>of</strong> layers<br />
and the number <strong>of</strong> units in each layer can change the complexity <strong>of</strong> the MLP. These networks<br />
are valuable tools in problems when one has little or no knowledge about the form <strong>of</strong> the<br />
relationship between input vectors and their corresponding outputs. The principle for the<br />
multi-layer perceptron neural network is that processing elements or nodes divided into layers<br />
perform calculations until an output value is computed at each <strong>of</strong> the output nodes (Foody<br />
1996). Data to the input layer are the <strong>satellite</strong> channel values used in the <strong>classification</strong><br />
procedure. Then one or more hidden layers perform the calculations and send the result to the<br />
output layer, which has one node for each class. Every node in a layer is connected to every<br />
node in the layer above and below, see figure 4. The connections carry weights, which<br />
encapsulate the behaviour <strong>of</strong> the network and are adjusted during training.<br />
11
Level<br />
k<br />
j<br />
i<br />
Connection<br />
weights<br />
Output Classes<br />
1 2<br />
Input Pattern Feature Values<br />
Figure 4. The Architecture <strong>of</strong> Multi-Layer Perceptron (<strong>based</strong> on Kanellopoulos 1997).<br />
For the hidden layers the input to each node is the sum <strong>of</strong> the scalar products <strong>of</strong> the<br />
incoming vector components with their respective weights,<br />
input<br />
∑<br />
= i<br />
where wji is the weight connecting node i to node j and outi is the output from node i.<br />
The output <strong>of</strong> a node j is,<br />
The function f denotes the activation function <strong>of</strong> each node. The most frequently used is the<br />
sigmoid activation function,<br />
where x =inputj. This ensures that the node acts like a thresholding device.<br />
The network learns using the back-propagation algorithm iteratively. This algorithm<br />
minimizes the mean square error between the network's output and the desired output,<br />
where P are all input patterns and dk is the desired output.<br />
j<br />
12<br />
w<br />
ji<br />
out<br />
out j = f ( input j )<br />
1<br />
f ( x)<br />
=<br />
1+<br />
exp( −x)<br />
1<br />
E = ∑∑( d k − out k )<br />
2P<br />
P k<br />
i<br />
2<br />
Output layer<br />
Hidden layer<br />
Input layer
It compares the calculated output with the desired one, and adjust the weights attached to<br />
the connections, until the difference between the outputs is reduced to an acceptable level and<br />
the set <strong>of</strong> weights is stable.<br />
The weights are from the beginning small and random and then adjusted by the<br />
generalised delta rule,<br />
where wkj (t+1) and wkj (t) are the weights connecting nodes k and j at iteration (t+1) and t<br />
respectively, η is a learning rate parameter.<br />
For nodes in the output layer,<br />
and for nodes in the hidden layers,<br />
wkj ( t + 1)<br />
= wkj<br />
( t)<br />
+ η(<br />
δ koutk<br />
)<br />
δ<br />
k<br />
= ( d k − outk<br />
) f ′ ( inputk<br />
)<br />
To measure the generalisation ability it is common to have a set <strong>of</strong> <strong>data</strong> to train the network<br />
and a separate set to assess the performance. Once the neural network has been trained, the<br />
weights are saved to be used in the <strong>classification</strong> phase.<br />
The activation level <strong>of</strong> an output unit is positively related to the strength <strong>of</strong> membership to<br />
the class associated with the unit and lies on a 0-1 scale. These levels are significantly<br />
correlated with the sub-pixel land cover composition and may be mapped as fraction images.<br />
The classifier encounters lot <strong>of</strong> problems to separate the classes when the classes have lot<br />
<strong>of</strong> similarities. A-priori knowledge <strong>of</strong> the relative class distribution could be used to apply<br />
non-random weights to the network inputs to further distinguish similar classes.<br />
Another disadvantage is that the back-propagation method is extremely slow. About<br />
thousand iterations need to be done before the network converges to a solution. Sometimes no<br />
solution is found (Augusteijn and Warrender 1998).<br />
3.7.2 Applications with Multilayer Perceptron Classifier<br />
Several applications have been performed with the multilayer perceptron classifier. Howald<br />
classified Landsat TM <strong>data</strong> into seven land cover classes using a three layer network, with a<br />
slightly better overall <strong>classification</strong> accuracy than a maximum likelihood classifier (1989).<br />
Kanellopoulos et al. used a four-layer neural network for classifying multi-temporal SPOT<br />
multi-spectral imagery into twenty land cover classes (1991). Owing to the complexity <strong>of</strong> the<br />
<strong>classification</strong>, a relatively large neural network architecture was required and therefore the<br />
computation time was quite long. Gualtieri and Withers used clustering <strong>of</strong> multispectral <strong>data</strong><br />
(1988). Ryan et al. (1991) reported the use <strong>of</strong> a neural network to categorise small blocks <strong>of</strong><br />
pixels to extract shoreline features from images.<br />
Foody and Boyd (1999) used a neural network to derive fuzzy <strong>classification</strong>s <strong>of</strong> land cover<br />
along a transect crossing the transition from moist semi-deciduous forest to savannah in West<br />
Africa in February and December 1990. They used NOAA AVHRR for the <strong>study</strong>. The<br />
training <strong>data</strong> was acquired from core regions <strong>of</strong> each class. The AVHRR provides five<br />
spectral channels, so the network had five input units. The architecture also consisted <strong>of</strong> five<br />
hidden units in a single layer and four output units and a logistic sigmoid transfer function<br />
was used. The initial network weights were set randomly between the limits <strong>of</strong> ± 0.5 and the<br />
learning rate and momentum were set to 0.1 and 0.9 respectively. When 1000 iterations <strong>of</strong><br />
13<br />
∑<br />
δ f ′ (<br />
input ) δ w<br />
j = k k kj
stochastic backpropagation learning algorithm had been performed, the RMS error was 0.096<br />
and 0.117 for the February and December <strong>data</strong> sets respectively. The method was able to<br />
represent both gradual and sharp changes in land cover, as those associated with the forestsavannah<br />
and savannah-water boundaries respectively. It also enabled important properties <strong>of</strong><br />
the boundaries to be inferred, such as that the width <strong>of</strong> the forest-savannah transitional zone<br />
remained relatively constant, but had migrated significantly, between the two time periods.<br />
3.7.2 Neural Network using Cascade-correlation<br />
Improvements have been made on back-propagation, since it is so slow. One <strong>of</strong> these<br />
improved architectures is the cascade-correlation (Augusteijn and Warrender 1998). Cascadecorrelation<br />
is a feed-forward neural network. The network builds its internal structure<br />
incrementally, during training. The initial network consists <strong>of</strong> only two layers: an input layer<br />
and an output layer, which are completely connected. These connections are trained until no<br />
more significant changes occur between iterations. If the total training error still is too high, a<br />
hidden node will be allocated and trained to further reduce this error. This goes on until the<br />
termination criterion is reached or until a maximum number <strong>of</strong> hidden nodes are reached and<br />
the conclusion is made that the network cannot learn the problem.<br />
3.7.3 Applications with Cascade-correlation<br />
A <strong>study</strong> made <strong>of</strong> Augusteijn and Warrender (1998) was conducted to investigate the ability <strong>of</strong><br />
a cascade-correlation neural network to delineate upland and forested wetland areas and to<br />
distinguish between different levels <strong>of</strong> wetness in a forested wetland. They used NASA’s<br />
Airborne Terrestrial Applications Sensor (ATLAS) multispectral <strong>data</strong> and Airborne Imaging<br />
Radar Synthetic Aperture Radar (AIRSAR) <strong>data</strong>. The input values used for training and<br />
testing <strong>of</strong> the neural network classifier were not simply digital number (DN) values indicating<br />
the reflectance <strong>of</strong> each pixel in a given band, but averages <strong>of</strong> DN values calculated over 3<br />
pixel by 3 pixel neighbourhoods inside the training and test polygons. The same sample size<br />
had to be used in each category for the network to learn all categories equally well. First a five<br />
class categorization was tried. But three <strong>of</strong> the five classes were too similar to be<br />
distinguishable, so they were merged together. Several networks were trained using different<br />
band combinations. Every network were always trained five times on the same <strong>data</strong> set, each<br />
time using a different selection <strong>of</strong> random initial values for the connection strengths, to reduce<br />
the influence <strong>of</strong> a specific set <strong>of</strong> initial values on network performance. Test sample<br />
agreement scored well over 80 per cent in most <strong>case</strong>s.<br />
3.7.4 Fuzzy ARTMAP<br />
ARTMAP achieves a synthesis <strong>of</strong> fuzzy logic and ART networks (Mannan et al. 1998). ART<br />
stands for Adaptive Resonance Theory. The architecture <strong>of</strong> fuzzy ARTMAP consists <strong>of</strong> four<br />
layers <strong>of</strong> neurons: the input-, category-, mapfield-, and output layer. The values <strong>of</strong> the spectral<br />
bands and their complement, <strong>of</strong> dimension n, are the input to the first layer, which consists <strong>of</strong><br />
2n neurons. The category layer starts with one neuron, but grows in number as the learning<br />
proceeds. The output and mapfield layers consist <strong>of</strong> as many neurons as there are classes.<br />
Two vigilance parameters ρ1 and ρ2 control the operation during learning and operational<br />
phases <strong>of</strong> the network.<br />
14
The learning phase is an iteratively process where for each input a category choice is<br />
calculated according to,<br />
A ∧W<br />
S =<br />
α + W<br />
where A is the input feature vector and W1 is the weight vector between the input layer and a<br />
node in the category layer.<br />
For the node which has the largest value <strong>of</strong> S the match ratio at mapfield is calculated by,<br />
where B is the output class vector and W2 is the weight vector between a chosen node <strong>of</strong><br />
category layer and the mapfield layer.<br />
If the Rm value is larger than ρ2 the weights are changed and a new input value is managed.<br />
The learning process <strong>of</strong> mapfield- and category layer weights is called resonance, and the<br />
weights are calculated by,<br />
( new)<br />
1<br />
If not, ρ1 = Rc for that input, where Rc is calculated by,<br />
1<br />
R m<br />
For the nodes which has Rc ≥ ρ1 it is checked if any has Rm ≥ ρ2, if so is the <strong>case</strong> the<br />
weights are changed and a new input is managed, otherwise a new nodes committed, the<br />
weights are changed and a new input managed. This goes on until all training samples are<br />
exhausted and the category layer nodes stop growing or the number <strong>of</strong> iterations exceeds T, a<br />
chosen positive constant.<br />
After this process a score S is calculated for each <strong>of</strong> the committed nodes. The node in the<br />
output layer, which corresponds to the node with the largest value <strong>of</strong> S, indicates the category<br />
<strong>of</strong> the input pixel.<br />
α can be chosen around 0.01, β can be set to 1.0 in the beginning <strong>of</strong> the learning phase and<br />
to a smaller value later. ρ1 and ρ2 are set to 1.0 for the most accurate results.<br />
There is an advantage compared to maximum-likelihood and the multilayer perceptron<br />
classifier (MLP) <strong>of</strong> about 5 % in the overall accuracy (Mannan et al. 1998). The method is<br />
stable, easy to use, with a smaller number <strong>of</strong> parameters to manage and faster than MLP.<br />
15<br />
1<br />
1<br />
B ∧W<br />
=<br />
B<br />
( old )<br />
1<br />
∗<br />
2<br />
W = β ( A ∧W<br />
) + ( 1−<br />
β ) W<br />
W = β ( A ∧W<br />
) + ( 1−<br />
β ) W<br />
( new)<br />
2<br />
2<br />
R c<br />
( old )<br />
2<br />
A ∧W1<br />
=<br />
A<br />
1<br />
2<br />
( old )<br />
1<br />
( old )<br />
2
3.7.5 Applications with Fuzzy ARTMAP<br />
Six multispectral images acquired by Linear Imaging Self-scanning Sensor (LISS-II) camera<br />
<strong>of</strong> Indian Remote Sensing Satellite (IRS-1B) have been used in an experiment in using fuzzy<br />
ARTMAP in <strong>classification</strong> by Mannan et al. (1998). The training samples were selected by<br />
visual interpretation <strong>of</strong> the scenes. In order to preserve the samples relative values, the grey<br />
levels were normalized. 40 per cent <strong>of</strong> the samples were used in the training phase and the rest<br />
were applied in the operational phase to assess the accuracy. The four spectral band pixel grey<br />
values and their complements were input to the network. In the output binary vector, the bit,<br />
which belongs to the input’s class, was set to 1 and the rest to 0:s.<br />
All weights were initially set to 1, but decreased as the learning progressed. The result was<br />
about 5 per cent better in the overall accuracy compared to multilayer perceptron- and<br />
maximum likelihood classifier. In this experiment 7483 pixels were correctly classified and<br />
397 misclassified.<br />
16
4. Data and methods<br />
4.1 Introduction<br />
This chapter describes the <strong>study</strong> area and the <strong>data</strong> that has been used in the <strong>classification</strong>.<br />
Two different <strong>s<strong>of</strong>t</strong>wares have been used to manage the problem and the method to do this, is<br />
described.<br />
4.2 Data<br />
4.2.1 Satellite <strong>data</strong><br />
For the <strong>case</strong> area, only one Indian IRS-1C image is available and that one is very cloudy.<br />
Since IRS-1C Wifs only have two spectral bands, at least two images are needed to get a<br />
reliable <strong>classification</strong>. So the choice has fallen on the Russian RESURS-O1 #3 with MSU-SK<br />
sensored images. Several RESURS images are available for the <strong>study</strong> area, but only three are<br />
more or less cloud-free. These are from July and August 1995 and September 1996. These<br />
RESURS images have five spectral bands, but only four can be acquired in parallell. Of these<br />
four spectral bands, only three are acceptably noise free to be used, these are the red band and<br />
the two near infrared bands, see figure 5. Only seven bands are allowed in the <strong>s<strong>of</strong>t</strong>ware, so the<br />
image from August 1995 is excluded.<br />
The <strong>study</strong> area constituted only a small part <strong>of</strong> each image and was therefore extracted to<br />
reduce the amount <strong>of</strong> <strong>data</strong>. The area was 22 km × 44 km.<br />
17<br />
Figure 5. RESURS multitemporal colour<br />
composite <strong>of</strong> the first near-infrared band,<br />
red band and the second near infrared<br />
band from July 1995 (RGB) over the <strong>study</strong><br />
area.
4.2.2 Reference <strong>data</strong><br />
As reference <strong>data</strong> for training and<br />
evaluation the existing land cover <strong>data</strong><br />
from the pilot production <strong>of</strong> the Swedish<br />
CORINE land cover over the Gripen area<br />
was used, see figure 6.<br />
CORINE stands for ‘Co-ordinating<br />
Information on the Environment’ and was<br />
a EU programme beginning 1985. This<br />
land cover is mainly <strong>based</strong> on information<br />
from digital topographical maps<br />
interpretation and <strong>classification</strong> <strong>of</strong> Landsat<br />
TM images with 30 m resolution,<br />
resampled to a pixelsize <strong>of</strong> 25 m.<br />
In most <strong>of</strong> Europe the CORINE land<br />
cover contains 44 classes, but Sweden has<br />
added 11 classes to better fulfil the needs<br />
(SSC 1998).<br />
The main classes are produced by: Figure 6. The reference image,<br />
Swedish CORINE land cover.<br />
♦ The wetland classes are mapped using masks from the digital topographic map and by<br />
classifying <strong>satellite</strong> <strong>data</strong>. The map provides the spatial limits and the <strong>satellite</strong> <strong>data</strong> is used<br />
to update the masks for wet mires and dry mires and for division <strong>of</strong> those classes. The<br />
mires near a lake or sea is by a GIS-operation named inland- or coastal marshes.<br />
♦ The water bodies are classified by thresholding in TM band 5, where that class has low<br />
digital values. Then the water is classified and subtracted from the water mask in the<br />
topographic map. The reason for this is that floating vegetation or lakes overgrown with<br />
reeds are obtained, and can be candidates for the inland marshes.<br />
♦ Forest is classified using codes from National Forest Survey field <strong>data</strong>. For given<br />
coordinates, these codes say what kind <strong>of</strong> forest that grows there. For the same<br />
coordinates, spectral signatures are extracted. By this the signatures for different kinds <strong>of</strong><br />
forest are known and the image can be classified under the forest mask. Filtered signatures<br />
are used.<br />
♦ Clear-cut forest is extracted from an image showing change, which is produced using<br />
histogram matching between the same channel in each scene. Filtered signatures are used<br />
for <strong>classification</strong> (nearest-neighbour).<br />
♦ Urban boundaries are interpreted on screen using the Statistics Sweden urban <strong>data</strong>base<br />
and <strong>satellite</strong> images as background. Buildings within 200 m are included.<br />
♦ Agricultural areas are taken from the digital topographical map, when it exists, as in my<br />
<strong>study</strong> area.<br />
The forest mask also contains the roads. These roads are mostly classified as forest<br />
regeneration or clear-cut forest. Narrow and long areas <strong>of</strong> these classes which coincides with<br />
18
oads narrower than 7 meters are alloted to the surrounding class. Then all areas in the image<br />
that were smaller than 2 pixels were removed.<br />
The received image was again resampled, now to 12.5 m.<br />
4.3 Study area<br />
4.3.1 General description <strong>of</strong> the Baltic Sea region landscape<br />
The Baltic Sea region is the area (i.e. countries) surrounding the Baltic Sea. From the<br />
vegetation’s point <strong>of</strong> view the countries are very young, in particularly Sweden, because <strong>of</strong><br />
the glaciation. The immigration <strong>of</strong> plants has occurred in several steps and from different<br />
directions (Bråkenhielm et al. 1998). The spreading <strong>of</strong> the plants is strongly dependent <strong>of</strong> the<br />
climate and their ability to compete with other species. The human activities, like agricultural<br />
and forestry have also affected the vegetation. Where a species chooses to grow is for<br />
example a matter <strong>of</strong> the bedrock, i.e. the content <strong>of</strong> nutrition in the mineral and rock and their<br />
ability to disintegrate and contain water. The disintegration give rise to the foundation <strong>of</strong> fine<br />
granular soils, which in turn helps keeping water, which in turn makes the chemical<br />
disintegration easier and results in more nutrition. The bedrock has developed from magmatic,<br />
sedimentary and metamorphic processes during billions <strong>of</strong> years. Different phases have<br />
contributed to bedrock <strong>of</strong> varying composition. This, along with agricultural fields with<br />
groups <strong>of</strong> trees and the variations in the terrain, makes the landscape patchy. How patchy a<br />
person or animal experience the landscape is due to the person’s references or the animals<br />
field <strong>of</strong> view or habitat. For example, doesn’t a hawk think the same area is patchy as do the<br />
mouse.<br />
This patchy landscape makes it difficult to generate a good and reliable hard <strong>classification</strong>,<br />
from medium resolution <strong>satellite</strong> images, since they consists <strong>of</strong> pixels with sizes between 150-<br />
250 m. It is not <strong>of</strong>ten such pixels contain just one class. Therefore, the main task for this<br />
project is to analyze if <strong>s<strong>of</strong>t</strong> classifiers can better show the real situation for such <strong>satellite</strong><br />
images.<br />
The <strong>study</strong> area is very suitable for this task, since it contains all kinds <strong>of</strong> land covers, from<br />
large agricultural fields with groves <strong>of</strong> trees, to large forests with patches <strong>of</strong> agricultural<br />
fields. In figure 7, it can be seen that there are several classes in a 150 m- pixel and therefore a<br />
hard classifier is not suitable.<br />
The <strong>study</strong> area, called Gripen, is situated between lake Hjälmaren and the southern part <strong>of</strong><br />
the Baltic Sea bay – Bråviken, in the southern part <strong>of</strong> Sweden. The area contains two big<br />
cities – Örebro and Norrköping. Around these cities, the landscape is very flat with large areas<br />
<strong>of</strong> arable land. The forested area between these cities is very patchy with many smaller lakes<br />
and hills. The forest contains both deciduous and coniferous forest and a mixture <strong>of</strong> both with<br />
no sharp boundaries.<br />
There are also different kinds <strong>of</strong> wetlands, which are <strong>of</strong> special interest in the BALANS<br />
project, since they act as filters <strong>of</strong> nutrients before the water runs into the Baltic Sea.<br />
19
Figure 7. A part <strong>of</strong> the <strong>study</strong> area in Gripen showing the reference image with 12.5 m pixel<br />
size and a 150 m grid to elicit the fuzziness in the medium resolution <strong>satellite</strong> image pixels<br />
(adapted from http://www.grida.no/baltic/htmls/maps.htm).<br />
4.3.2 Quantitative characterisation<br />
The area distribution <strong>of</strong> each class in the reference image are: water 5.19%, urban 1.82%,<br />
agricultural areas 20.08%, marsh 0.71%, deciduous forest 11.79%, coniferous forest 36.80%,<br />
mire 2.42%, clear-cut forest 7.35%, rock 0.82%, barren land 0.63%, peat 0.44% and grass<br />
3.51%. This result in 13% unknown classes when using eight classes and 9% unknown<br />
classes when using twelve classes.<br />
4.4 Tools<br />
4.4.1 IDRISI<br />
IDRISI is a geographic information and image processing <strong>s<strong>of</strong>t</strong>ware system developed by the<br />
Graduate School <strong>of</strong> Geography at Clark University, Worcester, U.S.A. It covers the full<br />
spectrum <strong>of</strong> GIS and remote sensing needs. The <strong>s<strong>of</strong>t</strong>ware has over 150 modules that provide<br />
facilities for the input, display and analysis <strong>of</strong> geographic <strong>data</strong>. For this project the fuzzy<br />
signature development and <strong>s<strong>of</strong>t</strong> classifier modules has been especially important. But at least<br />
this version, version 2.0 for Windows, has its limitations, which force the users to find other<br />
20
solutions to their problems, and sometimes import <strong>data</strong> from other systems, for example<br />
already radiometric- and geometric corrected images.<br />
The module <strong>based</strong> on the logic <strong>of</strong> Fuzzy Sets in IDRISI is FUZCLASS. In that module fuzzy<br />
set membership is determined from the distance <strong>of</strong> pixels from signature means. To use this<br />
module two parameters has to be set. The first is the z-score distance where fuzzy<br />
membership becomes zero. If a pixel has the same location as the class mean the membership<br />
grade is 1.0, and away from this point the grade decreases until it reaches zero. The second<br />
parameter is whether or not the membership values should be normalized. Normalization<br />
makes the assumption that the classes are the only one existing, and thus the membership<br />
values for all classes for a single pixel must sum to 1.0.<br />
The module using the Dempster-Shafer theory in IDRISI is BELCLASS. In normal use,<br />
Dempster-Shafer theory requires the classes under consideration to be mutually exclusive and<br />
exhaustive. But in BELCLASS the pixel may belong to some unknown class, for which a<br />
training site has not been provided. An additional category is added to every analysis called<br />
[other]. The result is consistent with Dempster-Shafer theory, but recognizes the possibility<br />
that there may be classes present about which we have no knowledge. The uncertainty is<br />
almost identical to that <strong>of</strong> Dempster-Shafer ignorance.<br />
In BELCLASS you have two choices <strong>of</strong> output: beliefs or plausibilities. In either <strong>case</strong>, one<br />
image <strong>of</strong> belief or plausibility is produced for each class. A <strong>classification</strong> uncertainty image is<br />
also produced.<br />
4.4.2 ERDAS IMAGINE<br />
ERDAS Imagine is a complete GIS and image processing <strong>s<strong>of</strong>t</strong>ware produced by the<br />
Engineering department <strong>of</strong> ERDAS, Inc., Atlanta, U.S.A. It is easy to use and has modules for<br />
most problems. The limitation in this <strong>case</strong> is that it only has one fuzzy module. For this<br />
project it has been used for preparation <strong>of</strong> the images and extraction <strong>of</strong> pixel information in<br />
the resulting images.<br />
21
4.5 Methodology<br />
The outline <strong>of</strong> the used method is shown in figure 8. A scheme showing the main action<br />
steps is included as appendix 1. The produced macro files are listed by name in a scheme in<br />
appendix 2.<br />
4.5.1 Outline<br />
Figure 8. The used methodology.<br />
4.5.2 Initial <strong>data</strong> preprocessing<br />
4.5.2.1 Reference <strong>data</strong><br />
Preprocessing<br />
reference image<br />
Binarisation,<br />
Thematic aggregation<br />
Spatial aggregation<br />
Sampling <strong>of</strong> training<br />
sites<br />
Preparation <strong>of</strong> training statistics<br />
Running the modules<br />
Evaluation and<br />
comparison <strong>of</strong> <strong>s<strong>of</strong>t</strong><br />
classifiers<br />
4.5.2.1.1 Binarisation<br />
The preprocessing <strong>of</strong> the reference image began with extraction <strong>of</strong> clouds that were in the<br />
RESURS images. This was done by using the mask made from the <strong>satellite</strong> images. To m<br />
make one image layer for each class a re<strong>classification</strong> was made. The result was a binary<br />
image for each class.<br />
22<br />
Preprocessing<br />
<strong>satellite</strong> images<br />
Radiometric correction<br />
Geometric correction
4.5.2.1.2 Thematic aggregation<br />
The reference image was reclassified to contain eight or twelve classes. These classes were<br />
water, urban, agricultural field (agri), marshes (marsh), deciduous forest (decid), coniferous<br />
forest (conif), wet/other mires (mire) and clear-cut forest (clear). The urban class consisted<br />
<strong>of</strong> the classes coarse- and dense city structures and smaller communities from the Swedish<br />
CORINE land cover. The coniferous and deciduous class consisted <strong>of</strong> conifers/decids with<br />
different age and height. When 12 classes were used, rock, barren land (barr), peat and<br />
grass were added.<br />
The choices were made <strong>based</strong> on the more detailed classes that the BALANS project<br />
wants to extract.<br />
The BALANS project has envisaged at least five broad classes.<br />
These are,<br />
• Artificial surfaces<br />
• Agricultural areas<br />
• Forests and semi-natural areas<br />
• Wetlands<br />
• Water bodies<br />
More detailed classes envisaged if possible is urban fabric, arable land, inland wetlands,<br />
coniferous- and deciduous forest. Therefore, a <strong>classification</strong> <strong>of</strong> these more detailed classes<br />
was tried out to.<br />
4.5.2.1.3 Spatial aggregation<br />
The reference image contained pixels with a size <strong>of</strong> 12.5 meter. To be able to compare the<br />
reference image with the RESURS images, which have a pixel size <strong>of</strong> 150 meter, an<br />
aggregation was made in IDRISI. In the process an average <strong>of</strong> twelve 12.5 m pixels in<br />
both the X- and Y-direction was calculated and was output as a 150 m pixel. That made<br />
the border pixels <strong>of</strong> each class patch fuzzy, i.e. they didn’t contain 100% <strong>of</strong> that class any<br />
more. One such layer were made for each class and these were seen as the ‘true’ fuzzy<br />
classes, since now the percent <strong>of</strong> fuzziness in each pixel was known.<br />
4.5.2.2 Satellite <strong>data</strong><br />
4.5.2.2.1 Radiometric correction<br />
Since RESURS images from two different dates were used they had different illumination.<br />
Therefore the image which visually seemed to be the best was set as reference in this <strong>case</strong>.<br />
Each band in the other image was histogram-matched against respectively band in the<br />
’reference’ image.<br />
4.5.2.2.2 Geometric correction<br />
The RESURS images didn’t have any known projection and didn’t lay above each other. In<br />
one image there was one band that didn’t lay above the other bands either. So the images<br />
were geometrically corrected image to image by taking ground control points (GCPs). The<br />
RMS error were about 25 m as maximum, that is about one sixth <strong>of</strong> a pixel. The images<br />
were then resampled with cubic convolution and then corrected against the reference image.<br />
The reference image was projected with Transverse Mercator and had datum and spheroid<br />
WGS 84, so after the correction against the reference image, the RESURS images received<br />
the same projection.<br />
A cloud mask was created by taken training sites in the clouds and other ‘close’ classes.<br />
A supervised <strong>classification</strong> with maximum-likelihood was performed and the output classes<br />
were tested to see which were actually clouds. These clouds were marked in an ‘area-<strong>of</strong>interest’-layer<br />
and cut out in all images.<br />
23
4.5.3 Choice <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />
The <strong>s<strong>of</strong>t</strong>ware available for <strong>s<strong>of</strong>t</strong> <strong>classification</strong> was IDRISI. This <strong>s<strong>of</strong>t</strong>ware included <strong>s<strong>of</strong>t</strong><br />
classifiers using Bayesian probability theory and Dempster-Shafer theory reviewed earlier. It<br />
also included one method where the distance from a class mean were set where the fuzzy<br />
membership <strong>of</strong> that class should be zero.<br />
4.5.4 Application <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />
4.5.4.1 Sampling <strong>of</strong> ’training <strong>data</strong>’ from reference <strong>data</strong><br />
Different training <strong>data</strong> samplings were examined.<br />
♦ The first sampling was to take random sample points in the aggregated masks, as training<br />
sites. In this sampling as well as in sampling two and three, different points were sampled<br />
for each class.<br />
♦ The second sampling was to make majority masks from the aggregated ones, were only<br />
the pixels with class values <strong>of</strong> 75-100 % (90-100% for water) were included, to receive<br />
purer sample points to see if there were any improvement.<br />
♦ The third sampling was to, first (in ERDAS Imagine) manually take training sites within<br />
the wetland mask to see where the wetland had a separable signature from the other<br />
classes. Then a mask from these wetland areas were made under the whole wetland layer<br />
and used as input to the aggregation in IDRISI. The sample points were taken in the<br />
majority masks.<br />
♦ The fourth sampling was to take random sample points in the whole image and divide<br />
them into eight equally quantified classes. In this <strong>case</strong> the points for a specific class could<br />
be sampled where that class didn’t have any pixel values. Most <strong>of</strong> the time the training<br />
samples for a specific class did not contain a majority <strong>of</strong> that class. All training samples<br />
were quite equal.<br />
4.5.4.2 Preparation <strong>of</strong> training statistics<br />
Each class’ training sites were seen as one training site group. After the sampling step, every<br />
training site group was compared to the aggregated masks to see how much <strong>of</strong> each class<br />
there was in every training site group. An average value was calculated for each class for each<br />
training site group, e.g. the output is an average value <strong>of</strong> content <strong>of</strong> water in the urban training<br />
site group and the content <strong>of</strong> water in the agricutural training site group etc. These average<br />
values were imported to the fuzzy partition matrix and linked to the image with the training<br />
site groups. The fuzzy partition matrix indicates the membership grades <strong>of</strong> each training site<br />
group in each class (Eastman 1997). After the linking to the training site groups image, one<br />
image for each class is created showing the membership grade in each training site group. The<br />
fuzzy signatures are created by giving each training site group a weight proportional to its<br />
membership grade when determining the mean, variance and covariance <strong>of</strong> each <strong>satellite</strong><br />
spectral band for each class.<br />
4.5.4.3 Running the modules<br />
After the creation <strong>of</strong> fuzzy signatures the different modules were tested. When the proportion<br />
<strong>of</strong> each class in the reference image is known, a prior probability can be input as a weight and<br />
improve the result.<br />
24
4.5.4 Evaluation and comparison <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />
4.5.4.1 Introduction to thematic accuracy assessment <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />
The ground <strong>data</strong> can rarely be assumed to be error free. So testing a <strong>classification</strong> is an<br />
evaluation <strong>of</strong> the level <strong>of</strong> agreement or correspondence between two sets <strong>of</strong> class allocations,<br />
both <strong>of</strong> which have their own error characteristics, rather than quantifying the accuracy.<br />
In hard <strong>classification</strong>, accuracy assessment is generally made with the use <strong>of</strong> an error<br />
matrix. Error matrices compare, on a class-by-class basis, the relationship between known<br />
reference <strong>data</strong> and the result <strong>of</strong> a <strong>classification</strong>. Such matrices are square, with as many rows<br />
and columns as there are classes to be assessed. Several characteristics are expressed by the<br />
error matrix, for example the error <strong>of</strong> omission (exclusion) and error <strong>of</strong> commission<br />
(inclusion). Then the producer’s accuracy, user’s accuracy and overall accuracy can be<br />
calculated. The producer’s- and user’s accuracy tells how much <strong>of</strong>, for example the<br />
coniferous class, that is correctly classified and how much <strong>of</strong> the classified category<br />
coniferous that is truly that in reality, respectively.<br />
Another measure to extract from the error matrix is the kappa statistic, which is a measure<br />
<strong>of</strong> the difference between the observed accuracy and a chance agreement between the<br />
reference <strong>data</strong> and a random classifier.<br />
The problem is that the error matrix is only appropriate for use with hard <strong>classification</strong><br />
(Foody 1996). In <strong>s<strong>of</strong>t</strong> <strong>classification</strong> you receive one matrix for each class, i.e. each pixel<br />
would be compared. But each pixel is just allowed to consist <strong>of</strong> one class, consequently to be<br />
correct or incorrect.<br />
A number <strong>of</strong> approaches have been suggested to assess the accuracy <strong>of</strong> <strong>s<strong>of</strong>t</strong> <strong>classification</strong>.<br />
For example, entropy can be used (Zhang and Foody 1998). Entropy describes the variations<br />
in class membership probabilities associated with each pixel.<br />
The entropy is calculated by,<br />
∑<br />
H ( p)<br />
= − p(<br />
x)<br />
log 2 p(<br />
x)<br />
where p(x) is the class membership probabilities.<br />
But entropy is only suitable when ground <strong>data</strong> are hard. When both classified <strong>data</strong> and<br />
ground <strong>data</strong> are fuzzy, cross-entropy are more appropriate (Zhang and Foody 1998). Other<br />
possible indices <strong>of</strong> <strong>classification</strong> accuracy may be <strong>based</strong> on correlation analysis and distance<br />
measures between the representations <strong>of</strong> the land cover in the image <strong>classification</strong> and ground<br />
<strong>data</strong>.<br />
One other alternative is to degrade the <strong>data</strong> to just contain one or two classes in each pixel<br />
and use conventional methods, but the result is not an evaluation <strong>of</strong> the fuzzy <strong>classification</strong>.<br />
There are a number <strong>of</strong> points to note. Firstly, the predictive accuracy <strong>of</strong> partial values tends<br />
to vary between the land cover classes being considered (Bastin 1997). Secondly, the<br />
estimates <strong>of</strong> pixel composition give no reliable estimate <strong>of</strong> where in the pixel objects or<br />
features are likely to be.<br />
To deal with the second point, there are methods for sharpening the fuzzy <strong>classification</strong><br />
output (Foody 1998). This is made by fusion between a fine spatial resolution image and the<br />
fuzzy <strong>classification</strong> derived at a coarser resolution.<br />
There is a difficulty in evaluating <strong>s<strong>of</strong>t</strong> <strong>classification</strong>s, and there need to be done more<br />
research for better methods. There is no general measure that says when the <strong>classification</strong> is<br />
good or bad.<br />
25
4.5.4.2 Criteria and methods available<br />
One method that can be used along the whole continuum <strong>of</strong> <strong>classification</strong> fuzziness, from the<br />
<strong>case</strong> where every stage is hard to the <strong>case</strong> where every stage is fuzzy, is the RMS error (rootmean-square<br />
error) between the estimated and actual class composition <strong>of</strong> the pixels (Foody<br />
1999),<br />
where xi is a measurement, t is the true value and n is the number <strong>of</strong> <strong>data</strong> to be assessed.<br />
4.5.4.3 A suggested new approach<br />
One new measure can be to calculate how many pixels that are under- or overclassified a<br />
certain percent, in 10%- intervals, for each module.<br />
4.5.4.4 Implementation<br />
RMS =<br />
4.5.4.4.1 Validation <strong>data</strong> from reference <strong>data</strong><br />
In the evaluation RMS errors were calculated for all samplings and modules to see<br />
how well the <strong>classification</strong> performed in the classes. The training site pixels and<br />
the ‘unknown’ areas in the reference image, i.e. classes that had not<br />
been used, were excluded and then the whole images were used in the validation. To be<br />
able to compare the modules and samplings, a weighted total average RMS error was<br />
calculated. Each class’ RMS error was weighted with the reference area <strong>of</strong> the same class,<br />
Average_RMS_error =Σ RMS_error⋅ percentuel area in the reference image<br />
Also the average value for each class, in percent was calculated. This was been done in<br />
two ways, one compared to the reference image and one compared to the classified pixels.<br />
This to be able to compare BELclass with BAYclass and the reference image. BELclass<br />
underclassifies and may just classify 10% <strong>of</strong> the image.<br />
An analysis were made to see if the <strong>classification</strong> improved when more training pixels<br />
were used, in the sampling where the pixels were randomly taken, using BELclass, and<br />
how many classes that could be extracted in the same sampling.<br />
4.5.4.4.2 Statistical analysis<br />
I used Micro<strong>s<strong>of</strong>t</strong> Excel for the calculations.<br />
( x t)<br />
∑ i −<br />
i<br />
n<br />
26<br />
2
5. Results<br />
5.1 Introduction<br />
This chapter contains the results in terms <strong>of</strong> RMS errors, over- and underclassified pixels and<br />
average values.<br />
5.2 Fuzzy signatures<br />
Some <strong>of</strong> the eight chosen classes extracted have very similar spectral signatures, when using<br />
two near-infrared bands and one red band. That is why the <strong>s<strong>of</strong>t</strong> classifiers classify so many<br />
pixels, with different percentage <strong>of</strong> probability, to wrong classes. The mean values for each<br />
spectral band for each class, for the sampling in majority masks, is shown in figure 9. There it<br />
can be seen that coniferous, clear-cut forest, deciduous forest, mire and urban have mean<br />
values close to each other and therefore the training pixels can be mixed.<br />
Mean value<br />
120<br />
100<br />
80<br />
60<br />
40<br />
20<br />
0<br />
-20<br />
Fuzzy signatures for the six spectral bands in the majority sampling<br />
1 2 3 4 5 6<br />
Spectral bands (NIR,NIR,NIR,NIR,red,red)<br />
Figure 9. The mean values for the fuzzy signatures in the six spectral bands used ( 1-2 near<br />
infrared from September 1996, 3-4 near infrared from July 1995, 5 red from July 1995 and 6<br />
red from September 1996).<br />
5.3 RMS errors<br />
To be able to compare the different modules and samplings, each class’ RMS error were<br />
weighted with the percentage area <strong>of</strong> the reference image for respective class, to retrieve an<br />
overall RMS error, see table 2 below. This table shows that BAYclass performs best<br />
<strong>classification</strong> result before FUZclass. BELclass is not that far behind. They differ just a little<br />
between the modules and samplings but the result improves when the training pixels get purer<br />
27<br />
agri<br />
clear<br />
conif<br />
decid<br />
marsh<br />
mire<br />
urban<br />
water
and separable signatures are used. The modules classified differently well for the different<br />
classes, as seen in the more detailed tables, 3-5, showing the RMS errors for each class,<br />
sampling and module. In table 3 and 4, it can be seen that the RMS errors are reduced for the<br />
wetland classes when using a mask with areas that manually had been tried out to be<br />
separable. A thematic class can not <strong>of</strong>ten be assembled in one spectral class. This is why<br />
clustering or supervised <strong>classification</strong> need to be done to achieve separable classes within the<br />
masks.<br />
The number <strong>of</strong> classes you try to extract in BELclass makes no difference in the RMS<br />
errors. This is because BELclass uses unknown classes. It is only the spectral signature that<br />
counts. In that way one class is independent to another class, see table 4. More classes in<br />
BAYclass on the other hand improves the RMS errors for most classes.<br />
Table 2. Overall RMS errors for eight classes, weighted with the area <strong>of</strong> each class in the<br />
reference image, for the different modules and methods (×100 for percent).<br />
Overall RMS for the different samplings, named 1-4:<br />
1. training sites within 2. within majority 3. in majority mask and with 4. randomly taken<br />
module aggregated mask masks wetland signatures trainingsites<br />
BAYclass 0.299 0.259 0.244 0.362<br />
BELclass 0.404 0.400 0.398 0.436<br />
FUZclass 0.304 0.283 0.258 0.373<br />
Table 3. RMS errors for BAYclass.<br />
RMS for 8 classes classified with BAYclass RMS for 12<br />
classes trainingsites<br />
within<br />
aggregated<br />
mask<br />
within majority<br />
masks<br />
28<br />
in majority mask randomly taken<br />
classes,<br />
BAYclass<br />
randomly<br />
and with<br />
training sites taken training<br />
wetland signatures sites<br />
water 0.13 0.11 0.11 0.23 0.18<br />
urban 0.16 0.15 0.15 0.19 0.13<br />
agri 0.23 0.23 0.22 0.35 0.4<br />
marsh 0.15 0.16 0.07 0.13 0.11<br />
decid 0.21 0.21 0.22 0.24 0.24<br />
conif 0.42 0.33 0.3 0.48 0.5<br />
mire 0.21 0.2 0.12 0.2 0.19<br />
clear 0.22 0.22 0.24 0.21 0.2<br />
rock 0.16<br />
barr 0.13<br />
peat 0.07<br />
grass 0.08
Table 4. RMS errors for BELclass.<br />
RMS for 8 classes classified with BELclass RMS for 12<br />
29<br />
classes<br />
classes training sites within majority in majority mask randomly randomly<br />
within aggregatedmasks and with wetlandtaken taken<br />
masks signatures training sites training sites<br />
water 0.19 0.20 0.20 0.23 0.22<br />
urban 0.11 0.11 0.11 0.13 0.13<br />
agri 0.41 0.41 0.41 0.46 0.46<br />
marsh 0.08 0.08 0.03 0.08 0.08<br />
decid 0.24 0.24 0.24 0.27 0.27<br />
conif 0.56 0.55 0.55 0.59 0.59<br />
mire 0.12 0.12 0.06 0.13 0.13<br />
clear 0.22 0.22 0.22 0.22 0.22<br />
rock 0.03<br />
barr 0.06<br />
peat 0.04<br />
grass 0.06<br />
Table 5. RMS errors for FUZclass.<br />
RMS for 8 classes classified with FUZclass<br />
classes training sites within within majority in majority mask and with randomly taken<br />
aggregated masks masks wetland signatures training sites<br />
water 0.13 0.11 0.11 0.22<br />
urban 0.13 0.13 0.15 0.18<br />
agri 0.22 0.21 0.18 0.37<br />
marsh 0.16 0.15 0.04 0.16<br />
decid 0.22 0.21 0.21 0.24<br />
conif 0.44 0.40 0.36 0.50<br />
mire 0.16 0.18 0.09 0.15<br />
clear 0.21 0.22 0.24 0.21<br />
5.4 Over- and under classified pixels<br />
When looking at the number <strong>of</strong> correct classified pixels (± 5% ) in the summaries <strong>of</strong> table 6<br />
and 7, it can be seen that BELclass is superior with an average <strong>of</strong> 80%, compared to<br />
BAYclass’ 63%. But in these tables, the values are not weighted with the area, and the small<br />
classes that don’t contain so many pixels therefore brings up the average value, since they are<br />
so well classified with BELclass. The under- and overestimated values for each class in the<br />
sum-up tables 6 and 7, is the percent in pixels, not the total RMS error. The detailed tables for<br />
each 10% interval are shown as tables 8 and 9.
Table 6. Percent pixels over- and underestimated, for BELclass (×100 for percent).<br />
Over-, correct- or underestimated for BELclass, method 2<br />
water urban agri marsh decid conif mire clear<br />
underestimated: 0.082 0.027 0.285 0.011 0.255 0.585 0.044 0.161<br />
correct +-5%: 0.918 0.968 0.708 0.970 0.702 0.409 0.929 0.808<br />
overestimated: 0.001 0.005 0.006 0.019 0.042 0.007 0.027 0.030<br />
average:<br />
underest. 0.181<br />
correct 0.802<br />
overest. 0.017<br />
Table 7. Percent pixels over- and underestimated pixels for BAYclass (×100 for percent).<br />
Over-, correct- or underestimated for BAYclass, method 2<br />
water urban agri marsh decid conif mire clear<br />
underestimated: 0.026 0.014 0.140 0.009 0.187 0.486 0.029 0.109<br />
correct +- 5%: 0.907 0.848 0.820 0.823 0.522 0.354 0.423 0.341<br />
overestimated: 0.067 0.139 0.040 0.169 0.291 0.160 0.548 0.550<br />
average:<br />
underest. 0.125<br />
correct 0.630<br />
overest. 0.245<br />
BELclass systematically underclassifies, as seen in table 8 or sum-up table 6. Figure 10 and<br />
11 visualizes the clear-cut forest class from table 8 and 9 to easier see how many pixels that<br />
are over- and underclassifed for BAYclass and BELclass and how much.<br />
30
Table 8. Percent pixels over- and underclassified for BELclass. For example 5.1% <strong>of</strong><br />
agricultural areas are underestimated between 84-95%.<br />
RMS in 10 % intervals for BELclass, method 2<br />
water urban agri marsh decid Conif mire clear<br />
100-94% underestimated 0.012 0.004 0.070 0.002 0.004 0.099 0.004 0.010<br />
95-84% 0.0130.0030.051 0.001 0.008 0.096 0.003 0.009<br />
85-74% 0.009 0.002 0.036 0.001 0.013 0.077 0.002 0.011<br />
75-64% 0.008 0.0030.030 0.001 0.020 0.067 0.004 0.013<br />
65-54% 0.009 0.0030.025 0.001 0.027 0.064 0.004 0.017<br />
55-44% 0.007 0.002 0.020 0.001 0.027 0.047 0.004 0.014<br />
45-34% 0.006 0.002 0.013 0.001 0.024 0.030 0.003 0.012<br />
35-24% 0.004 0.002 0.012 0.001 0.031 0.032 0.004 0.017<br />
25-14% 0.005 0.002 0.012 0.001 0.040 0.031 0.006 0.022<br />
15-4% 0.009 0.002 0.016 0.002 0.060 0.0430.010 0.037<br />
5% underest. - 5% overest. 0.918 0.968 0.708 0.970 0.702 0.409 0.929 0.808<br />
6-15% overestimated 0 0.0030.004 0.009 0.025 0.004 0.021 0.016<br />
16-25% 0 0.001 0.001 0.004 0.010 0.002 0.005 0.007<br />
26-35% 0 0 0 0.003 0.004 0 0.001 0.003<br />
36-45% 0 0 0 0.001 0.002 0 0 0.001<br />
46-55% 0 0 0 0.001 0.001 0 0 0.001<br />
56-65% 0 0 0 0 0 0 0 0.001<br />
66-75% 0 0 0 0 0 0 0 0<br />
76-85% 0 0 0 0 0 0 0 0<br />
86-95% 0 0 0 0 0 0 0 0<br />
96-100% 0 0 0 0 0 0 0 0<br />
Procent pixlar<br />
Procent över- eller underklassade pixlar i hyggesklassen, klassad med BELclass<br />
0,900<br />
0,800<br />
0,700<br />
0,600<br />
0,500<br />
0,400<br />
0,300<br />
0,200<br />
0,100<br />
0,000<br />
100-94% underestimated<br />
95-84%<br />
85-74%<br />
75-64%<br />
65-54%<br />
55-44%<br />
45-34%<br />
35-24%<br />
25-14%<br />
15-4%<br />
5% underest. - 5% overest.<br />
6-15% overestimated<br />
16-25%<br />
26-35%<br />
36-45%<br />
46-55%<br />
56-65%<br />
66-75%<br />
76-85%<br />
86-95%<br />
96-100%<br />
10 % - intervall för över- och underklassat<br />
Figure 10. Percent under- and overclassified pixels for the clear-cut forest class with<br />
BELclass.<br />
31<br />
hyggen
Percent pixels<br />
Table 9. Percent pixels over- and underclassified for BAYclass. For example 1.4 % <strong>of</strong><br />
agricultural areas are underestimated between 84-95%.<br />
Over- or underclassified, in 10 % intervals for BAYclass, sampling 2<br />
water urban agri marsh decid conif mire clear<br />
100-94% underestimated 0.001 0 0.012 0 0 0.001 0 0<br />
95-84% 0.001 0.001 0.014 0 0.002 0.005 0.001 0.002<br />
85-74% 0.001 0.001 0.012 0 0.004 0.014 0.001 0.005<br />
75-64% 0.002 0.001 0.0130 0.009 0.035 0.0030.010<br />
65-54% 0.0030.002 0.014 0.001 0.015 0.060 0.003 0.012<br />
55-44% 0.0030.001 0.013 0.001 0.022 0.081 0.004 0.015<br />
45-34% 0.002 0.001 0.011 0.001 0.025 0.083 0.004 0.015<br />
35-24% 0.003 0.002 0.013 0.001 0.031 0.080 0.004 0.015<br />
25-14% 0.004 0.002 0.015 0.001 0.036 0.069 0.004 0.016<br />
15-4% 0.007 0.0030.023 0.0030.044 0.058 0.005 0.019<br />
5% underest. - 5% overest. 0.907 0.848 0.820 0.823 0.522 0.354 0.423 0.341<br />
6-15% overestimated 0.036 0.081 0.014 0.087 0.152 0.074 0.217 0.172<br />
16-25% 0.011 0.018 0.009 0.025 0.070 0.0430.1730.226<br />
26-35% 0.008 0.009 0.006 0.013 0.038 0.022 0.094 0.121<br />
36-45% 0.005 0.006 0.005 0.009 0.018 0.011 0.036 0.026<br />
46-55% 0.002 0.005 0.002 0.008 0.009 0.005 0.014 0.004<br />
56-65% 0.002 0.004 0.002 0.005 0.004 0.0030.006 0<br />
66-75% 0.001 0.0030.001 0.005 0.001 0 0.003 0<br />
76-85% 0.001 0.0030.001 0.006 0 0 0 0<br />
86-95% 0 0.005 0.001 0.007 0 0 0 0<br />
96-100% 0 0.005 0 0.004 0 0 0 0<br />
1.000<br />
0.900<br />
0.800<br />
0.700<br />
0.600<br />
0.500<br />
0.400<br />
0.300<br />
0.200<br />
0.100<br />
0.000<br />
Percent pixels under- or overclassified in the 'clear' class with BAYclass<br />
100-94% underestimated<br />
95-84%<br />
85-74%<br />
75-64%<br />
65-54%<br />
55-44%<br />
45-34%<br />
35-24%<br />
25-14%<br />
15-4%<br />
5% underest. - 5% overest.<br />
6-15% overestimated<br />
16-25%<br />
26-35%<br />
36-45%<br />
46-55%<br />
56-65%<br />
66-75%<br />
76-85%<br />
86-95%<br />
96-100%<br />
Percent under- or overclassified intervals<br />
Figure 11. Percent under- and overclassified pixels for the clear-cut forest class with<br />
BAYclass.<br />
32<br />
clear
5.5 Average values<br />
If you look at the average value, in percent <strong>of</strong> the reference image, for each class classified<br />
with BELclass, in table 10, you see that no sampling performs very well. This can be due to a<br />
low amount <strong>of</strong> training sites. It is a limitation in the <strong>s<strong>of</strong>t</strong>ware that only allow 10 000 random<br />
points. The small classes need <strong>of</strong> up to 40 000 points to get about 100 training pixels for each<br />
class in the aggregated mask. So probably the result improves if the training sites are collected<br />
manually and separable signatures are extracted. BAYclass have some higher occurence and<br />
comes therefore closer to the true average value for some classes, but overclassifies for many<br />
classes as well. It is not so strange that BAYclass have higher occurence, since this module<br />
think that the trained classes are the only one existing. If there are other classes in the image,<br />
it puts them to the most look-alike class. BELclass on the other hand treats the pixel as<br />
belonging to an unknown class if the pixel doesn’t support any <strong>of</strong> the given classes.<br />
As seen in average table 11, BELclass classifies more <strong>of</strong> the image when more classes are<br />
added, 13% compared to 0.90% for eight classes. But it is in wrong classes the increase is<br />
done.<br />
33
Table 10. Average values for 8 classes in percent <strong>of</strong> reference image and percent <strong>of</strong> classified<br />
image.<br />
Average values for 8 classes (in %) Average values for<br />
classified with BELclass for the different samplings BAYclass and FUZclass<br />
for one sampling<br />
class reference training sites within in majority mask randomly BAYclass FUZclass<br />
image within % <strong>of</strong> majority % <strong>of</strong> and with wetland % <strong>of</strong> taken % <strong>of</strong> in majority in majority<br />
aggregated classified mask, classified signatures, classified training classified mask, mask,<br />
mask, area % <strong>of</strong> ref. area % <strong>of</strong> ref. area sites area % <strong>of</strong> ref. % <strong>of</strong> ref.<br />
% <strong>of</strong> ref.<br />
% <strong>of</strong> ref.<br />
water 6.03 0.98 7.88 0.78 6.90 0.75 6.44 0.03 2.8 5.81 6.16<br />
urban 2.11 0.37 2.95 0.35 3.10 0.35 3.03 0.38 42.6 5.71 9.23<br />
agri 23.3 4.04 32.57 2.98 26.46 3.15 27.21 0.00 0.4 15.47 17.03<br />
marsh 0.82 0.82 6.65 0.46 4.12 0.01 0.08 0.36 39.9 6.35 7.40<br />
decid 13.68 2.37 19.11 2.23 19.77 2.47 21.33 0.00 0.3 12.74 12.35<br />
conif 42.71 2.84 22.93 3.07 27.23 3.18 27.47 0.00 0.0 25.18 18.30<br />
mire 2.81 0.46 3.75 0.46 4.13 0.06 0.51 0.06 6.9 13.54 14.25<br />
clear 8.53 0.52 4.16 0.93 8.29 1.61 13.94 0.06 7.1 15.20 15.28<br />
Sum %: 100.00 12.40 100.00 11.26 100.00 11.58 100.00 0.90 100.0 100.00 100.00<br />
35
Table 11. Average values for 12 classes in percent <strong>of</strong> reference image and percent <strong>of</strong><br />
classified image.<br />
Average values (in %) for 12 classes<br />
classified with BELclass and BAYclass for sampling 4<br />
class reference BELclass,<br />
BAYclass,<br />
image random training sites<br />
random training sites<br />
% <strong>of</strong> ref. % <strong>of</strong> classified % <strong>of</strong> classified area<br />
area<br />
water 5.67 0.16 1.16 12.07<br />
urban 1.99 0.01 0.04 5.98<br />
agri 21.93 0.00 0.00 5.68<br />
marsh 0.77 7.81 56.84 5.23<br />
decid 12.87 0.00 0.00 7.09<br />
conif 40.19 0.00 0.01 10.39<br />
mire 2.65 0.00 0.02 13.86<br />
clear 8.03 0.01 0.08 8.72<br />
rock 0.89 0.02 0.12 13.61<br />
barr 0.69 3.47 25.22 6.62<br />
peat 0.48 1.43 10.42 4.97<br />
grass 3.83 0.84 6.08 5.80<br />
Sum %: 100.00 13.74 100.00 100.00<br />
As can be seen in table 12, the number <strong>of</strong> training pixels does not matter when they are<br />
taken randomly. That is because the fuzzy partitioning matrix in the <strong>s<strong>of</strong>t</strong>ware doesn’t allow<br />
a not-square matrix and therefore an average value is calculated for each class.<br />
Table 12. Comparison <strong>of</strong> RMS errors with 3000<br />
training sample points and 10 000.<br />
RMS for 8 classes classified with BELclass<br />
classes randomly taken randomly taken<br />
trainingsites trainingsites<br />
RESURS 3000 RESURS 10000<br />
water 0.23 0.23<br />
urban 0.13 0.13<br />
agri 0.46 0.46<br />
marsh 0.08 0.08<br />
decid 0.27 0.27<br />
conif 0.59 0.59<br />
mire 0.13 0.13<br />
clear 0.22 0.22<br />
In figure 12-37 the reference images for all classes along with the result for each class for<br />
BAYclass and BELclass are shown for visual comparison. The agricultural class, the<br />
deciduous-, coniferous and clear-cut forest classes are labelled with shortenings in the figure<br />
texts.<br />
36
Figure 12. The agri.class<br />
from the reference image.<br />
Figure 15. The water class<br />
from the reference image.<br />
Figure 18. The urban class<br />
from the reference image.<br />
Figure 13. The agri.class<br />
classified with BAYclass.<br />
Figure 16. The water class<br />
classified with BAYclass.<br />
Figure 19. The urban class<br />
classified with BAYclass.<br />
37<br />
Figure 14. The agri.class<br />
classified with BELclass.<br />
Figure 17. The water class<br />
classified with BELclass.<br />
Figure 20. The urban class<br />
classified with BELclass.
Figure 21. The mire class<br />
from the reference image.<br />
Figure 24. The decid class<br />
from the reference image.<br />
Figure 27. The clear class<br />
from the reference image.<br />
Figure 22. The mire class<br />
classified with BAYclass.<br />
Figure 25. The decid class<br />
classified with BAYclass.<br />
Figure 28. The clear class<br />
classified with BAYclass.<br />
38<br />
Figure 23. The mire class<br />
classified with BELclass.<br />
Figure 26. The decid class<br />
classified with BELclass.<br />
Figure 29. The clear class<br />
classified with BELclass.
Figure 30. The conif class<br />
form the reference image.<br />
Figure 33. The marsh class<br />
from the reference image.<br />
Figure 31. The conif class<br />
classified with BAYclass.<br />
Figure 34. The marsh class<br />
classified with BAYclass.<br />
Figure 36. Uncertainty<br />
image for BAYclass.<br />
39<br />
Figure 32. The conif class<br />
classified with BELclass.<br />
Figure 35. The marsh class<br />
classified with BELclass.<br />
Figure 37. Uncertainty<br />
image for BELclass.
6. Discussion<br />
The reason why the water is surprisingly bad classified in all tryouts, can be due to error in<br />
the reference image. One lake in the reference image is practically covered with reed in the<br />
RESURS images. Training pixels in that lake can reduce the training site average for the<br />
class, and make at least BELclass to not classify water pixels if to far away from the mean<br />
value. BAYclass handles the problem a bit better since it only chooses among the classes<br />
given. Coniferous forest and agricultural areas have the highest RMS errors. This can be due<br />
to the fact that these classes are not spectrally homogenous. Coniferous pixels can also have<br />
been classified as water, since water and coniferous are the darkest pixels in the image. One<br />
class that is surprisingly well classified is the urban class. This can be due to that time series<br />
are used.<br />
There are many reasons why the classifiers perform unsatisfactory. The major reason is<br />
the input <strong>data</strong>. The RESURS images only had three, or actually only two spectral bands with<br />
good enough radiometric quality. These were the two near infrared and the red bands.<br />
Therefore, two images from different times were used, to at least have six bands. But in one<br />
<strong>of</strong> these images the spectral bands did not fit the other so a geometric correction had to be<br />
done. All images were then geometrically corrected to each other and to the reference image.<br />
There can still be small errors in the geometry, which affects the <strong>classification</strong> and the<br />
evaluation.<br />
The reference <strong>data</strong> is not really true either. It has been developed using topographic maps to<br />
get the boundaries <strong>of</strong> the classes and then for some classes just classified within the mask.<br />
Some other classes have been extracted using GIS-operations.<br />
The reference image is <strong>based</strong> on interpretation, GIS-operations and maximum-likelihood<br />
<strong>classification</strong>, why the true percentage distribution is not really known. What is really<br />
between 50-100% is set to be 100%. So the assumptions about what really is 100% may not<br />
be true. The classes, which are mixed in the reference image, as for example mixed forest, are<br />
not considered in the validation. The content <strong>of</strong> each forest kind in a pixel is unknown. But<br />
they have not participated in the training phase either.<br />
40
7. Conclusions and recommendations<br />
The advantages <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers are that small classes will not vanish as with maximumlikelihood<br />
<strong>classification</strong> and they gives a measure, not in whole pixels, <strong>of</strong> the occurrence <strong>of</strong><br />
the classes.<br />
To receive an acceptable <strong>classification</strong> result, the training areas need to be spectrally<br />
separable. This can be done with clustering or expert knowledge. There is also needed to be<br />
collected enough number <strong>of</strong> training areas to get the spectral variation <strong>of</strong> each class.<br />
If only red and near infrared bands are available, a time series is required to get enough<br />
information. The images can be chosen close in time to get rid <strong>of</strong> clouds and more important<br />
from different vegetation periods for a better <strong>classification</strong>. The <strong>satellite</strong> images also needs to<br />
have good geometric and radiometric quality.<br />
Forest and agricultural areas are hardest to classify because <strong>of</strong> the heterogeneous spectral<br />
signatures. Urban on the other hand is surprisingly well classified. This can be due to the time<br />
series. BAYclass is the <strong>s<strong>of</strong>t</strong> classifier that gives the best result.<br />
There need to be done some improvements in the <strong>s<strong>of</strong>t</strong>ware to be able to use any <strong>of</strong> these<br />
<strong>s<strong>of</strong>t</strong> classifiers in production. The main drawback is that the fuzzy partition matrix has to be<br />
square. So even if you take training sites with separable signatures manually, you have to<br />
calculate an average value for all your training sites for a certain class. This reduces the value<br />
<strong>of</strong> many training sites.<br />
The second drawback, if still wanting the computer to take training sites, is that not enough<br />
training sites can be taken for small classes. Not enough randomly spread out pixels will hit<br />
the areas where a member <strong>of</strong> a small class is situated.<br />
The third drawback if wanting to use images from different dates, is that the maximum<br />
number <strong>of</strong> spectral bands allowed is seven when using IDRISI.<br />
When these improvements are done, these <strong>s<strong>of</strong>t</strong> classifiers can be a valuable tool when using<br />
medium- or coarse resolution <strong>satellite</strong> <strong>data</strong> to extract more information from the images.<br />
41
References<br />
Arora, M.K. and Foody, G.M. 1997, Log-linear modelling for the evaluation <strong>of</strong> the variables<br />
affecting the accuracy <strong>of</strong> probabilistic, fuzzy and neural network <strong>classification</strong>.<br />
International Journal <strong>of</strong> Remote Sensing 4:785-798<br />
Augusteijn, M.F. and Warrender, C.E. 1998, Wetland <strong>classification</strong> using optical and radar<br />
<strong>data</strong> and neural network <strong>classification</strong>. International Journal <strong>of</strong> Remote Sensing 8:1545-<br />
1560<br />
Bastin, L. 1997, Comparison <strong>of</strong> fuzzy c-means <strong>classification</strong>, linear mixture modelling and<br />
MLC probabilities as tools for unmixing coarse pixels. International Journal <strong>of</strong> Remote<br />
Sensing 17:3629-3648<br />
Bråkenhielm, S., Drakenberg, B., Hellgren, M., Holmgren, P., Karltun, E., Melkerud, P-A.<br />
and Olsson, M. 1998, Markinfo. Institutionen för skoglig marklära, SLU http://wwwmarkinfo.slu.se<br />
(visited October the 15 th 1999)<br />
Eastman, J.R. 1997, IDRISI for Windows Version 2.0: User´s Guide. Clark Labs, Clark<br />
University, Worcester, Main, USA<br />
euromap Satellitendaten- Vertriebsgesellschaft mbH, 1997, IRS-1C Handbook.<br />
http://www.euromap.de/ (visited October the 10 th 1999)<br />
Foody, G.M. 1996, Approaches for the production and evaluation <strong>of</strong> fuzzy land cover<br />
<strong>classification</strong> from remotely-sensed <strong>data</strong>. International Journal <strong>of</strong> Remote Sensing 7:1317-<br />
1340<br />
Foody, G.M. 1998, Sharpening fuzzy <strong>classification</strong> output to refine the representation <strong>of</strong><br />
sub-pixel land cover distribution. International Journal <strong>of</strong> Remote Sensing 13:2593-2599<br />
Foody, G.M. 1999, The continuum <strong>of</strong> Classification Fuzziness in Thematic Mapping.<br />
Photogrammetric Engineering & Remote Sensing 4:443-451<br />
Foody, G.M. and Boyd, D.S. 1999 Fuzzy mapping <strong>of</strong> tropical land cover along an<br />
environmental gradient from remotely sensed <strong>data</strong> with an artificial neural network.<br />
Journal <strong>of</strong> Geographical Systems 1:23-35<br />
Fuller, D. and Fulk, M. 1998, An Assessment <strong>of</strong> Fire Distribution and Impacts during 1997<br />
in Kalimantan, Indonesia Using Satellite Remote Sensing and Geographic Information<br />
Systems. The George Washington University and Clark University<br />
Gualtieri, J.A. and Withers, J. 1988, Goddard researchers simulate neural networks using<br />
parallel processing. NASA Information Systems Newsletter 15:12-15<br />
Howald, K.J. 1989, Neural network image <strong>classification</strong>. Proceedings ASPRS-ACSM Fall<br />
Convention from Compass to Computer, Cleveland 207-215<br />
Hyyppä, J. 1999, WP 3200 State <strong>of</strong> the art in EO mapping, monitoring and <strong>data</strong> accessibility:<br />
Report on Earth Observation methods and <strong>data</strong> accessibility <strong>of</strong> mediumresolution<br />
<strong>satellite</strong> <strong>data</strong>. BALANS – Planning and management in the Baltic Sea Region<br />
with land information from EO, Environment and Climate Programme, Area 3.3: Centre for<br />
Earth Observation.<br />
Kanellopoulos, I., Varfis, A., Wilkinson, G.G. and Mégier, J. 1991, Classification <strong>of</strong><br />
Remotely-Sensed Satellite Images Using Multi-Layer Perceptron Networks in Proceedings<br />
<strong>of</strong> the 1991 International Conference on Artificial Neural Networks (ICANN-91), Espoo,<br />
Finland (Kohonen, T., Mäkisara, K., Simula, O. and Kangas, J.,eds.) Elsevier Science<br />
Publications (North-Holland) 2:1067-1070<br />
42
Kanellopoulos, I. 1997, Use <strong>of</strong> Neural Networks for Improving Satellite Image Processing<br />
Techniques for Land Cover/Land Use Classification, SUPCOM'95 LOT 16. Provision <strong>of</strong><br />
Statistical Services – Support to the Commission EUROSTAT Remote Sensing and<br />
Statistics Programme. Institute for Systems, Informatics and Safety (Neural Networks<br />
Laboratory) & Space Applications Institute (EMAP Unit - Advanced Methods Sector),<br />
European Commission, Joint Research Centre I-21020 Ispra (VA), Italy<br />
http://ams.egeo.sai.jrc.it/eurostat/Lot16-SUPCOM95/ (visited September the 15 th 1999)<br />
Langaas, S. 1996, Transboundary European GIS <strong>data</strong>bases: A review <strong>of</strong> the Baltic Sea<br />
region experiences, GISDATA Specialist Meeting on Geographic Information: the<br />
European Dimension. BALANS – Planning and management in the Baltic Sea Region with<br />
land information from EO, Environment and Climate Programme, Area 3.3: Centre for<br />
Earth Observation.<br />
Lillesand, T.M. and Kiefer, R.W. 1994, Remote Sensing and Image Interpretation. John<br />
Wiley & Sons, Inc. 587-618<br />
Malmberg, U. and Furberg, O. 1997 Klassnings- och kompositmetoder för medelupplösande<br />
satellit<strong>data</strong>, Pilotstudie med RESURS-<strong>data</strong> över centrala Östersjöområdet. Swedish Space<br />
Corporation, Rymdbolaget X-PUBL-11<br />
Mannan, B. and Roy, J. and Ray, A.K. 1998, Fuzzy ARTMAP supervised <strong>classification</strong> <strong>of</strong><br />
multi-spectral remotely-sensed images. International Journal <strong>of</strong> Remote Sensing 4:767-774<br />
Manyara, C.G. and Lein, J.K. 1994, Exploring the Suitability <strong>of</strong> Fuzzy Set Theory in Image<br />
Classification: A Comparative Study Applied to the Mau Forest Area Kenya.<br />
ASPRS/ACSM, http://www.odyssey.maine.edu/gisweb/spatdb/acsm/ac94044.htm<br />
(visited September the 16 th 1999)<br />
Maselli, F. And Rodolfi, A. and Conese, C. 1996, Fuzzy <strong>classification</strong> <strong>of</strong> spatially degraded<br />
Thematic Mapper <strong>data</strong> for the estimation <strong>of</strong> sub-pixel components. International Journal<br />
<strong>of</strong> Remote Sensing 3:537-551<br />
Olsson, B. 1999, BALANS Land Cover Information for the Baltic Sea Drainage Basin.<br />
BALANS – Planning and management in the Baltic Sea Region with land information from<br />
EO, http://www.ssc.se/rst/rss/BALANS/ (visited October th 15 th 1999)<br />
Palubinskas,G. 1994, Post-processing <strong>of</strong> fuzzy <strong>classification</strong> <strong>of</strong> forests for Landsat TM<br />
imagery. EGIS Foundation<br />
Pouncey, R. and Flynn, T. and Curry, K. 1997, ERDAS IMAGINE Ver. 8.3.1 On-Line<br />
Manuals. ERDAS, Inc.<br />
Ryan, T.W., Sementelli, P.J., Yuen, P. and Hunt, B.R. 1991, Extraction <strong>of</strong> shoreline features<br />
by neural nets and image processing. Photogrammetric Engineering and Remote Sensing<br />
7:947-955<br />
SSC Rymdbolaget, 1998, Produktspecifikation av landtäcke<strong>data</strong> version 1.0. SSC<br />
SSC Satellitbild, 1999, RESURS-O1 Technical Specifications.,<br />
http://www.ssc.se/sb/RESURS/technic.html (visited October the 11 th 1999)<br />
Zhang, J. and Foody, G.M. 1998, A fuzzy <strong>classification</strong> <strong>of</strong> sub-urban land cover from<br />
remotely sensed imagery. International Journal <strong>of</strong> Remote Sensing 14:2721-2738<br />
43
Appendix 1: Action steps for <strong>s<strong>of</strong>t</strong> <strong>classification</strong><br />
This is an example for water.<br />
Actions Input Module Output<br />
Make a water mask reference_gripen RECLASS rcwater<br />
in 12.5 m pixelsize<br />
Aggregate to a pixel rcwater CONTRACT aggwater<br />
size <strong>of</strong> 150 m<br />
Scale the pixel values aggwater SCALAR scwater<br />
to be in range 1-100,<br />
but still real values<br />
Convert the pixel values scwater CONVERT cvwater<br />
to byte binary<br />
For method 2 and 3:<br />
Mask out 75-100 % <strong>of</strong> cvwater RECLASS majwater<br />
a class, 90-100% for cvdecid… majdecid…<br />
water<br />
Convert the majority majwater CONVERT cv2water<br />
masks to integer binary majdecid… cv2decid…<br />
Take out sample points majwater SAMPLE vecwater<br />
randomly to be majdecid… vecdecid…<br />
training sites from the<br />
mask.<br />
Initialize images with aggwater INITIAL trswater<br />
the same parameters as trsdecid…<br />
an aggregated image<br />
The initialized image vecwater POINTRAS trswater<br />
updated with rasterized vecdecid… trsdecid…<br />
sample points<br />
Reclass the sample trswater RECLASS rc2water<br />
points to 1 and 0 trsdecid… rc2decid…<br />
Overlay to extract just majwater, rc2water OVERLAY ovwater<br />
the sample points which majdecid, rc2decid… ovdecid…<br />
are within the mask<br />
44
Numbering the sample ovwater GROUP grwater<br />
points ovdecid… grdecid…<br />
Reclass the sample grwater RECLASS grwater1<br />
points for one class to grdecid… grdecid1…<br />
only one classnumber<br />
Make an 8-layer image, grwater1 OVERLAY gr1_8<br />
(if you have 8 classes), grdecid1…<br />
with the classes (sample<br />
point groups).<br />
Extract an average value grwater1, aggwater, EXTRACT valwater<br />
for each class in each grdecid1, aggdecid… valurban…<br />
class mask.<br />
Create a fuzzy partitioning matrix. In Database Workshop<br />
Number 8 rows for the use <strong>of</strong> 8 classes. -Number the rows. One more<br />
Name 8 columns with class names, row comes up with the down<br />
e.g water, decidious. arrow. To create one more<br />
column:<br />
-Modify<br />
-Add field<br />
Real (4 byte) and class name<br />
Put in the values in the valwater In Database Workshop<br />
matrix valurban… -File<br />
-import/export Idridi values file<br />
free format ASCII (VAL file)<br />
valwater… and corresponding<br />
class<br />
Link the matrix to the gr1_8, water In Database Workshop fzwater<br />
training sites (sample gr1_8, decid… -Assign field values to fzdecid….<br />
points groups) image gr1_8 fzwater..<br />
and corresponding class<br />
Create the fuzzy gr1_8, sig.names, FUZSIG water.sig<br />
signatures <strong>satellite</strong>bands decid.sig….<br />
Classify with different signature names MAXLIKE mlclass<br />
hard/<strong>s<strong>of</strong>t</strong> classifiers BAYCLASS baywater…<br />
BELCLASS belwater…<br />
FUZCLASS fuzwater…<br />
45
Calculate the over- baywater, aggwater OVERLAY errwater<br />
and underclassified baydecid, aggdecid… errdecid…<br />
pixel values<br />
Add 1 to the extracted errwater SCALAR sc2water<br />
error-values and errdecid… sc2decid…<br />
multiply with 100 to<br />
have positive values<br />
between 1 and 200<br />
Convert the pixel values sc2water CONVERT cv3water<br />
to byte binary sc2decid… cv3decid…<br />
Export the error images and <strong>classification</strong> images to Imagine to receive pixel information:<br />
Export the images to cv3water, baywater -File expwater.img<br />
ERDAS Imagine cv3decid… - Export ex2water.img<br />
-Export<br />
-S<strong>of</strong>tware-Specific Formats<br />
-ERDIDRIS<br />
-IDRISI to ERDAS 7.4<br />
name <strong>of</strong> file to export and name in<br />
Imagine with extension .img<br />
In Imagine:<br />
-Open the image<br />
-info i<br />
-compute statistics<br />
In the viewer:<br />
-Raster<br />
-Attributes<br />
-Edit<br />
Mark the column<br />
-Copy<br />
Copy the pixel information into excel to calculate RMS and average values for each class<br />
and each module.<br />
In Excel:<br />
-Paste the number <strong>of</strong> pixels in the excel-file “mall_RMS.xls” and “mall_medeltal.xls”<br />
46
Appendix 2: Macr<strong>of</strong>ile- scheme<br />
8 classes, aggregated mask<br />
NEWMASK<br />
8 classes, majority mask<br />
MACMAJ_3<br />
MACMAJ_5<br />
MACR_8_2<br />
TRSITES3<br />
MACR8_4<br />
MCAGG8_5<br />
8 classes, miresig, aggregated mask<br />
NEWMASK2<br />
MACR8_6 , create matrix<br />
FUZSIG (FUZSIG4 if Landsat TM)<br />
MAC8_BAY MAC8_BEL MAC8_FUZ MAC8_ML<br />
Export to Imagine and Excel. eval.. for RMS for random<br />
cv10.. for average for random<br />
cv4.. for RMS other methods<br />
cv3.. for average other methods<br />
47<br />
8 classes, random<br />
MAC_RD_2 , create matrix<br />
FUZSIG2 (FUZSIG3 IF Landsat TM)
TRITA-GEOFOTO serien utges av:<br />
Kungl Tekniska Högskolan<br />
Institutionen för Geodesi och Fotogrammetri<br />
100 44 STOCKHOLM<br />
Drottning Kristinas väg 30<br />
Telefon: 08-7907342<br />
Telefax: 08-7907343<br />
URL: http://www.geomatics.kth.se/<br />
TRITA-GEOFOTO-serien 1999 - The TRITA-GEOFOTO series, 1999<br />
1999:1 Friborg, Klas. Differential GPS in A Tested Mobil Environment – An Accuracy Study.<br />
March 1999. (Examensarbete i geodesi nr 3061. Handledare: Horemuz)<br />
1999:2 Torlegård, Kennert. Analytisk fotogrammetri och dess felteori. (Kompendium)<br />
1999:3 Horemuz, Milan and Lars E Sjöberg. Quick GPS Ambiguity Resolution For Short<br />
And Long Baselines. March 1999.<br />
1999:4 Anna-Maria Dorfh och Maria Nilsson. Digitalt ort<strong>of</strong>oto för ajourhållning av<br />
baskarta. April 1999. (Examensarbete i fotogrammetri nr 112. Handledare: Anders<br />
Boberg)<br />
1999:5 Rose-Marie Kvål och Annica Larsson. Algal Bloom Monitoring and Detection in the<br />
Baltic Sea, the Usefulness <strong>of</strong> Marine Sensors Compared to a Meteorological Sensor. April<br />
1999. (Examensarbete i fotogrammetri. Handledare: Boberg, Lundén och Rud)<br />
1999:6 Krzyszt<strong>of</strong> Gajdamowicz. Automated Processing <strong>of</strong> Georeferenced Colour Stereo Images<br />
for Road Inventory. April 1999. (Doctoral dissertation)<br />
1999:7 Daniel Öhman. Automatic Text Localisation on Traffic Signs from Colour Stereoscopic<br />
Image Sequences. April 1999. (Examensarbete i fotogrammetri nr 113. Handledare: Boberg,<br />
Gajdamowicz)<br />
1999:8 Stefan Wulcan. Samlingskarta - En förstudie avseende samordning av ledningar inom<br />
Solna. May 1999. (Examensarbete. Handledare: Hauska)<br />
1999:9 Magnus Johansson och Christer Bylund. Landhöjning i södra Sverige<br />
beräknad från tredje precisionsavvägningen och RH70 höjder. Juni 1999. (Examensarbete i<br />
geodesi nr 3062. Handledare: Fan)<br />
1999:10 Jan Danielsen. Determination <strong>of</strong> Land Uplift in areas not covered by repeated levellings.<br />
With application to South-Norway. Juni 1999. (Licentiate thesis)<br />
1999:11 Mattias Pettersson. Utvärdering av vektoriseringsprogram. Juni 1999.<br />
(Examensarbete i geoinformatik. Handledare: Hauska)<br />
1999:12 Anders Vidén. Semiautomatiserad fotogrammetrisk mätning av signalerade punkter, linjeoch<br />
ellipsobjekt i MIDOC, Multi Image Documentation.". Juni 1999. (Examensarbete i<br />
fotogrammetri nr 114. Handledare: Axelsson)<br />
1999:13 Albert Morales Pérez and Pere Carmona Ortega. Movements monitoring <strong>of</strong> a high<br />
building. Juni 1999. (Examensarbete i geodesi nr 3063. Handledare: Horemuz)
1999:14 Lina Nyberg och Andreas Prevodnik. A Pilot Study <strong>of</strong> Soil Erosion in the Close<br />
Environment <strong>of</strong> Idrija, Western Slovenia. An Application <strong>of</strong> Geographical Information<br />
Systems. December 1999. (Examensarbete i geoinformatik. Handledare: Henkel)<br />
1999:15 Sanna Sparr Olivier. Évaluation de la possibilité d'utiliser des images<br />
<strong>satellite</strong>s à haute résolution et le GPS différentiel pour la cartographie en<br />
haute montagne - Avec réalisation d'un prototype. Augusti 1999.<br />
(Examensarbete i fotogrammetri nr 115. Handledare: Boberg)<br />
1999:16 Rannala, Marek. Surveying <strong>of</strong> road geometry with Real Time Kinematics GPS. Oktober<br />
1999. (Examensarbete i geodesi nr 3064. Handledare: Horemuz)<br />
1999:17 Klang, Dan. Reconstruction <strong>of</strong> Geometric Road Data Using Remotely Sensed Imagery.<br />
Oktober 1999. Photogrammetric Reports No 67. (Doctoral Dissertation)<br />
TRITA-GEOFOTO-serien 2000 - The TRITA-GEOFOTO series, 2000<br />
2000:1 Carballo, Gabriel. Statistically-<strong>based</strong> multiresolution network flow phase unwrapping in<br />
SAR interferometry. January 2000. Photogrammetric Reports No 68. (Doctoral Dissertation)<br />
2000:2 Jonas Andersson och Jens Hedlund. Geografiska analyser över Internet. Januari 2000.<br />
(Examensarbete i geoinformatik. Handledare: Roslund)<br />
2000:3 Thomas Rehders. Noggrannhetsstudier vid RTK-mätning och kvalitetsundersökning av GPSmottagare.<br />
Februari 2000. (Examensarbete i geodesi nr 3065. Handledare: Horemuz)<br />
2000:4 Åsa Hesson. Publicering av historiska kartor på Internet. Februari 2000. (Examensarbete i<br />
geoinformatik. Handledare: Hauska)<br />
2000:5 Malin Skagerström och Karin Wiklander. Extension <strong>of</strong> a control network in dar es salaam<br />
using GPS-technology. Februari 2000. (Examensarbete i geodesi nr 3066. Handledare:<br />
Stoimenov)<br />
2000:6 Maria Sjödin och Anna Strid. Automatisk generalisering med objektorienterad kart<strong>data</strong>bas.<br />
Februari 2000. (Examensarbete i geoinformatik. Handledare: Hauska)<br />
2000:7 Per Koldestam. IT-projektet Ortnamnslinjen. Mars 2000. (Examensarbete i geoinformatik.<br />
Handledare: Hauska).<br />
2000:8 Anna Källander och Elin Söderström. Deformationsanalys i horisontal- och vertikalled på<br />
Slussen-konstruktionen. Mars 2000. (Examensarbete i geodesi nr 3067. Handledare: Horemuz<br />
och Lewen)<br />
2000:9 Anna Haglund. <strong>Towards</strong> <strong>s<strong>of</strong>t</strong> <strong>classification</strong> <strong>of</strong> <strong>satellite</strong> <strong>data</strong> - A <strong>case</strong> <strong>study</strong> <strong>based</strong> upon Resurs<br />
MSU-SK <strong>satellite</strong> <strong>data</strong> and land cover <strong>classification</strong> within the Baltic Sea Region. Mars 2000.<br />
(Examensarbete i geoinformatik. Handledare: Roslund och Langaas).