13.01.2013 Views

Towards soft classification of satellite data A case study based ... - KTH

Towards soft classification of satellite data A case study based ... - KTH

Towards soft classification of satellite data A case study based ... - KTH

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Towards</strong> <strong>s<strong>of</strong>t</strong> <strong>classification</strong> <strong>of</strong> <strong>satellite</strong> <strong>data</strong><br />

A <strong>case</strong> <strong>study</strong> <strong>based</strong> upon Resurs MSU-SK <strong>satellite</strong> <strong>data</strong> and land<br />

cover <strong>classification</strong> within the Baltic Sea Region.<br />

Anna Haglund<br />

Master <strong>of</strong> Science Project no. In Geoinformatics<br />

Royal Institute <strong>of</strong> Technology<br />

Department <strong>of</strong> Geodesy and Photogrammetry<br />

Stockholm, Sweden<br />

March 2000


Universitetsservice US AB<br />

Stockholm 2000<br />

08-790 7400


Preface<br />

This master <strong>of</strong> science project has been performed at the department <strong>of</strong> geodesy and<br />

photogrammetry, division for geoinformatics by assignment <strong>of</strong> Satellus, during the autumn<br />

1999. The project comprises 20 credits and is the completion <strong>of</strong> the education for technical<br />

surveyors at the Surveying Program at the Royal Institute <strong>of</strong> Technology, <strong>KTH</strong>, in Stockholm.<br />

I would like to thank my supervisors at <strong>KTH</strong>: Sindre Langaas, dept.<strong>of</strong> Civil and<br />

Environmental Engineering and Maria Roslund, dept. <strong>of</strong> Geodesy and Photogrammetry.<br />

I would also like to thank Satellus in Solna, where I have performed the project.<br />

Especially I want to thank Kjell Wester, who has been my supervisor at Satellus and has<br />

given many technical advises and come up with ideas.<br />

I<br />

Stockholm, mars 2000<br />

Anna Haglund


Sammanfattning<br />

Vid användning av medelupplösande satellitbilder med en pixelstorlek på 150-250 m på<br />

marken kan det vara svårt att välja homogena träningsytor. Pixelstorleken är också större än<br />

många objekt på marken, varför de kanske försvinner i en ”hård klassning”, där varje pixel<br />

bara kan tilldelas en klass. Ibland vill man att varje pixel ska innehålla så många klasser som<br />

det är i verkligheten. För att klara dessa problem har s k. mjuka klassificerare utvecklats.<br />

Syftet med rapporten är att undersöka hur många klasser som kan fås från<br />

medelupplösande satellit<strong>data</strong> när man använder högupplösande <strong>data</strong> som referens för<br />

generering av träningsytor och fastställa hur bra klassificeringen blir.<br />

Undersökningen gjordes med delar av två RESURS-bilder från två olika tidpunkter, med<br />

tre spektrala band, över en del av södra Sverige. Som referens användes svenska<br />

landtäcke<strong>data</strong> (SLD). För fyra olika samplingsmetoder jämfördes resultatet från IDRISIs<br />

moduler för mjuka klassificerare med varandra. Det skattade medelfelet varierade mellan<br />

klasserna och var minst för de små klasserna och förekomsten för varje klass jämfört med<br />

referensbilden skiljde sig relativt mycket. En orsak till den bristfälliga överensstämmelsen kan<br />

delvis bero på att referensbilden inte är helt korrekt. Referensbilden är baserad på information<br />

från den digitala topografiska kartan, varifrån klassgränser tagits och klassning <strong>of</strong>tast skett<br />

bara innanför gränsen med maximum-likelihood. Med antagandet att referensbilden är korrekt<br />

kommer felaktigheten i referens<strong>data</strong> att skyllas på RESURS-klassningen och därmed försämra<br />

den bedömda noggrannheten. Passningen mellan RESURS-scenerna och referensbilden är<br />

också en felkälla liksom passningen RESURS-scenerna sinsemellan. Så det finns både<br />

tematiska och geometriska fel. En annan orsak är för få träningsytor tillsammans med att<br />

klasserna inte alltid är spektralt separerbara.<br />

En förbättring av mjukvaran måste också göras innan någon användning i produktion kan<br />

bli aktuell.<br />

III


Abstract<br />

When using medium resolution <strong>satellite</strong> images with a pixel size <strong>of</strong> 150-250 m on the ground<br />

it can be difficult to collect homogenous training sites. The pixel size is also larger than many<br />

objects on the ground, why they might disappear in a “hard <strong>classification</strong>”, where every pixel<br />

just can contain one class. The wish is <strong>of</strong>ten that every pixel should contain as many classes<br />

as there are in reality. To manage these problems, so called <strong>s<strong>of</strong>t</strong> classifiers have been<br />

developed.<br />

The purpose <strong>of</strong> the paper is to examine how many classes that can be extracted from<br />

medium resolution <strong>satellite</strong> <strong>data</strong> when using higher resolution <strong>satellite</strong> <strong>data</strong> as reference and<br />

how well the <strong>classification</strong> performs.<br />

This examination was conducted using parts <strong>of</strong> two RESURS images from two different<br />

dates, with three spectral bands, covering a section <strong>of</strong> the southern parts <strong>of</strong> Sweden. As<br />

reference the national land cover <strong>classification</strong> was used. For four different samplings I<br />

compared the results from the IDRISI modules for <strong>s<strong>of</strong>t</strong> <strong>classification</strong> with each other. The<br />

RMS errors varied between the classes and were smallest for the small classes and the<br />

occurrence in the image for each class compared to the reference image differed relatively<br />

much. One reason for the insufficient agreement can partly be due to that the reference image<br />

is not completely correct. The reference image is <strong>based</strong> on information from the digital<br />

topographical map, from which class boundaries has been taken and the <strong>classification</strong> mostly<br />

been performed within these, with maximum-likelihood. With the assumption that the<br />

reference image is correct, the errors in this will be blamed on the RESURS <strong>classification</strong> and<br />

thus worsen the assessed accuracy. The geometric correction between the RESURS images<br />

and the reference image, as well as the geometric correction between the RESURS images<br />

them selves, are also an error source. So there are both thematic and geometric errors. One<br />

other reason is too few training sites along with not always spectrally separable classes.<br />

There need to be done some improvements in the <strong>s<strong>of</strong>t</strong>ware before any use in production can<br />

be <strong>of</strong> interest.<br />

IV


Contents<br />

PREFACE...................................................................................................................................................................I<br />

SAMMANFATTNING.......................................................................................................................................... III<br />

ABSTRACT.............................................................................................................................................................IV<br />

CONTENTS ............................................................................................................................................................. V<br />

1. INTRODUCTION ................................................................................................................................................1<br />

1.1BACKGROUND...................................................................................................................................................1<br />

1.2 OBJECTIVES......................................................................................................................................................1<br />

1.3 THE BALANS PROJECT ..................................................................................................................................1<br />

1.4 STRUCTURE OF THE THESIS..............................................................................................................................2<br />

2. MEDIUM RESOLUTION SENSORS...............................................................................................................3<br />

2.1 AVAILABLE SENSORS .......................................................................................................................................3<br />

2.2 FUTURE SENSORS .............................................................................................................................................4<br />

3. A REVIEW OF SOFT CLASSIFIERS..............................................................................................................5<br />

3.1 INTRODUCTION.................................................................................................................................................5<br />

3.2 FUZZY SET THEORY.........................................................................................................................................6<br />

3.2.1 Fuzzy c-means (k-means) ........................................................................................................................7<br />

3.2.2 Applications with Fuzzy c-means classifier............................................................................................7<br />

3.3 SOFTENING MAXIMUM LIKELIHOOD CLASSIFIER (MCL).................................................................................8<br />

3.3.1 Applications with <strong>s<strong>of</strong>t</strong>ening MCL ...........................................................................................................8<br />

3.4 LINEAR MIXTURE MODELLING.........................................................................................................................8<br />

3.4.1 Applications with linear mixture modelling ...........................................................................................9<br />

3.5 METHODS BASED ON BAYESIAN PROBABILITY THEORY .................................................................................9<br />

3.5.1 Applications with Bayesian probability theory ....................................................................................10<br />

3.6 METHODS BASED ON DEMPSTER-SHAFER THEORY ......................................................................................10<br />

3.6.1 Applications with Dempster-Shafer theory ..........................................................................................11<br />

3.7 NEURAL NETWORK........................................................................................................................................11<br />

3.7.1 The Multilayer Perceptron Classifier using feed-forward and back-propagation.............................11<br />

3.7.2 Applications with Multilayer Perceptron Classifier ............................................................................13<br />

3.7.2 Neural Network using Cascade-correlation.........................................................................................14<br />

3.7.3 Applications with Cascade-correlation................................................................................................14<br />

3.7.4 Fuzzy ARTMAP......................................................................................................................................14<br />

3.7.5 Applications with Fuzzy ARTMAP........................................................................................................16<br />

4. DATA AND METHODS ...................................................................................................................................17<br />

4.1 INTRODUCTION...............................................................................................................................................17<br />

4.2 DATA ..............................................................................................................................................................17<br />

4.2.1 Satellite <strong>data</strong> ..........................................................................................................................................17<br />

4.2.2 Reference <strong>data</strong> .......................................................................................................................................18<br />

4.3 STUDY AREA...................................................................................................................................................19<br />

4.3.1 General description <strong>of</strong> the Baltic Sea region landscape......................................................................19<br />

4.3.2 Quantitative characterisation ...............................................................................................................20<br />

4.4 TOOLS.............................................................................................................................................................20<br />

4.4.1 IDRISI ....................................................................................................................................................20<br />

4.4.2 ERDAS IMAGINE .................................................................................................................................21<br />

4.5 METHODOLOGY..............................................................................................................................................22<br />

4.5.1 Outline....................................................................................................................................................22<br />

4.5.2 Initial <strong>data</strong> preprocessing .....................................................................................................................22<br />

4.5.3 Choice <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers .......................................................................................................................24<br />

V


4.5.4 Application <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers................................................................................................................24<br />

4.5.4 Evaluation and comparison <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers .....................................................................................25<br />

5. RESULTS ............................................................................................................................................................27<br />

5.1 INTRODUCTION...............................................................................................................................................27<br />

5.2 FUZZY SIGNATURES .......................................................................................................................................27<br />

5.3 RMS ERRORS .................................................................................................................................................27<br />

5.4 OVER- AND UNDER CLASSIFIED PIXELS .........................................................................................................29<br />

5.5 AVERAGE VALUES..........................................................................................................................................33<br />

6. DISCUSSION......................................................................................................................................................40<br />

7. CONCLUSIONS AND RECOMMENDATIONS..........................................................................................41<br />

REFERENCES.........................................................................................................................................................42<br />

APPENDIX 1: ACTION STEPS FOR SOFT CLASSIFICATION...............................................................................44<br />

APPENDIX 2: MACROFILE- SCHEME ..............................................................................................................47<br />

VI


1. Introduction<br />

1.1 Background<br />

Satellite images have been used for land cover <strong>classification</strong> ever since the first earth resource<br />

<strong>satellite</strong>s were launched. The most used method for <strong>classification</strong> has been the maximumlikelihood<br />

algorithm. Recently, medium resolution <strong>satellite</strong> <strong>data</strong> (150-250 m) has begun to be<br />

provided. Also then was the maximum-likelihood algorithm used in <strong>classification</strong>. There have<br />

been several studies done with the maximum-likelihood classifier. One made with medium<br />

resolution <strong>satellite</strong> <strong>data</strong> was done by Malmberg and Furberg (1997). The last five years there<br />

has been a progress on the method side. New classifiers have been developed – <strong>s<strong>of</strong>t</strong><br />

classifiers. They have been developed as a consequence that coarse- and medium resolution<br />

<strong>satellite</strong> <strong>data</strong> <strong>of</strong>ten contain mixed pixels, e.g. forests have no sharp boundaries between<br />

coniferous- and deciduous forest.<br />

It is a new and exciting field. These new <strong>s<strong>of</strong>t</strong> classifiers can extract more information than<br />

the maximum-likelihood classifier.<br />

1.2 Objectives<br />

The main objective is to evaluate and compare the suitability <strong>of</strong> some <strong>s<strong>of</strong>t</strong> classifiers to<br />

deduce land cover information in a patchy landscape using medium resolution <strong>satellite</strong> <strong>data</strong>,<br />

in particular for a <strong>study</strong> area in the Baltic Sea region for the purpose <strong>of</strong> the BALANS project<br />

(Landcover Information for the Baltic Sea Drainage Basin).<br />

Secondary objectives are to provide information about present and future <strong>satellite</strong>s with<br />

sensors that provide medium resolution <strong>data</strong> and make a review <strong>of</strong> the available <strong>s<strong>of</strong>t</strong><br />

classifiers today and there requirements on input- and reference <strong>data</strong>.<br />

1.3 The BALANS Project<br />

The Remote Sensing Technology Division at Satellus, a subsidiary to the Swedish Space<br />

Corporation, has in partnership with six organisations in Sweden, Norway, Finland and<br />

Poland, started a project called BALANS. The BALANS project’s aim is to develop a<br />

<strong>satellite</strong>-<strong>based</strong> land cover <strong>data</strong>base for the complete Baltic Sea Drainage Basin, see figure 1,<br />

using medium resolution <strong>satellite</strong>s (Olsson 1999). Many organisations in the Baltic Sea region<br />

have a common need for basic geographical <strong>data</strong> sets. Up to now the main need has been from<br />

the environmental, hydrometeorological and marine sciences, but now there are growing<br />

requirements from the physical planning, political and business communities (Langaas 1996).<br />

Today, there are many growing pressures on the Baltic Sea. There is a critical need for<br />

development that is sustainable, both economically and environmentally.<br />

The <strong>data</strong>base is envisaged to have a maximum resolution in the order <strong>of</strong> 200 meters.<br />

The project will last for three years, and will be finished sometime in the end <strong>of</strong> year 2001<br />

or the beginning <strong>of</strong> year 2002.<br />

1


Figure 1. The Baltic Drainage Basin and the BALANS <strong>study</strong> area<br />

(adapted from http://www.grida.no/baltic/htmls/maps.htm).<br />

1.4 Structure <strong>of</strong> the thesis<br />

In chapter 2, a review <strong>of</strong> present and future <strong>satellite</strong>s with medium resolution <strong>satellite</strong> <strong>data</strong> is<br />

provided. Chapter 3 is a review <strong>of</strong> available <strong>s<strong>of</strong>t</strong> classifiers and some applications made with<br />

those. Chapter 4 contains the <strong>data</strong> and method used in this <strong>study</strong>, chapter 5 the result and<br />

discussion, followed by chapter 6 with the conclusions.<br />

2


2. Medium Resolution Sensors<br />

2.1 Available sensors<br />

Medium resolution sensors are sensors that have a pixel size covering the ground <strong>of</strong> about<br />

150-250 meter. They bridge the gap in image extent and spatial resolution between<br />

SPOT/Landsat TM (20-30 m) and NOAA AVHRR (1.1 km) and are appropriate for mapping<br />

<strong>of</strong> large areas, since not so many images are needed and the images are not too coarse.<br />

Two <strong>satellite</strong>s with medium resolution sensors covering the Nordic area are the Russian<br />

<strong>satellite</strong> RESURS-O1 and the Indian Remote Sensing <strong>satellite</strong>, IRS. The <strong>satellite</strong>s have<br />

different sensors. RESURS-O1 has a MSU-SK sensor and IRS has a wide field sensor (WIFS),<br />

see descriptions and <strong>data</strong> in table 1. The only available MSU-SK <strong>data</strong> today is archived<br />

RESURS-O1 #3 <strong>data</strong> (SSC Satellitbild 1999). One orbit period takes 98 minutes and the<br />

repetition cycle for that <strong>satellite</strong> is 21 days, with a potential coverage <strong>of</strong> a specific area every<br />

third to fourth day at the equator. Every scene covers 600 x 600 km 2 . It is a wide-angle<br />

instrument with a conical scan where every pixel has the same size and viewing angle. The<br />

<strong>data</strong> from the curved scanning lines are resampled by cubic convolution from the nominal 170<br />

meters raw pixels to 160 meter pixels to obtain a quadratic image grid.<br />

The WIFS has two spectral bands, the red and the near infrared, with the applications to see<br />

chlorophyll absorption <strong>of</strong> plants and to survey biomass and see waterbodies, respectively<br />

(euromap 1997). One orbit around the earth takes 101.35 minutes to complete and it takes 341<br />

orbits to cover the entire earth. The repetition cycle for the IRS <strong>satellite</strong>s is 24 days, but IRS<br />

1C/D works as a system so their combined repetition is 12 days. At the Nordic latitudes a<br />

specific area is covered every 1-2 days. Also the WIFS’s scanning lines are resampled by cubic<br />

convolution to obtain a quadratic image grid.<br />

Both <strong>satellite</strong> types have large swath width. The largest difference between the <strong>satellite</strong>s is<br />

the number <strong>of</strong> wavelength bands. While RESURS-O1 MSU-SK has five bands: red, green, two<br />

near infrared (NIR) and one thermal infrared (TIR), IRS WIFS only has two: the red and near<br />

infrared. But in practice the TIR band in MSU-SK cannot be acquired in parallel with the<br />

others. Experiences from the <strong>data</strong> indicates that there are better radiometric dynamics in <strong>data</strong><br />

from WIFS on IRS than in the <strong>data</strong> from MSU-SK on RESURS-O1 #3 (Hyyppä 1999).<br />

3


Table 1. Summary <strong>of</strong> <strong>satellite</strong> and sensor <strong>data</strong>.<br />

Satellites:<br />

RESURS-O1 MSU-SK IRS-1C WiFS IRS-1D WiFS<br />

Orbit sun-synchronous, circular sun-synchronous, circular, near polar<br />

Average altitude 678 km 817 km<br />

Inclination 98.04 deg 98.69 deg<br />

Orbit period 98 min 101.35 min<br />

Orbit repeat cycle 21 days 12 days (24 days for each)<br />

Launch date 4/11 1994 28/12 1995 29/12 1997<br />

Sensors:<br />

Imaging mechanism Conical scan 2048 element linear array CCD<br />

Viewing angle 39 deg ?<br />

Spectralband: Wavelength: pixelsize: Wavelength: pixelsize:<br />

1: Green 0.5-0.6 µm 160 m -<br />

2: Red 0.6-0.7 µm 160 m 0.62-0.68 µm 188 m<br />

3:NIR 0.7-0.8 µm 160 m 0.77-0.86 µm 188 m<br />

4: NIR 0.8-1.1 µm 160 m -<br />

5: TIR 10.4-12.6 µm 600 m -<br />

Swath width 600 km 810 km 728-812 km<br />

Potential coverage <strong>of</strong> same area 3-4 days at the equator 1-2 days at Nordic latitudes<br />

Overlap ? Ca. 80%<br />

2.2 Future sensors<br />

There are two new medium resolution sensors coming. These are MODIS on the EOS AM-1<br />

and MERIS on the ENVISAT <strong>satellite</strong> (Hyyppä 1999).<br />

MODIS has a cross-track scan mirror and a set <strong>of</strong> linear arrays with spectral interference<br />

filters located in four focal planes. It has a viewing swath width <strong>of</strong> 2330 km and will cover the<br />

earth every 1-2 days. It will have 36 spectral bands in the range <strong>of</strong> 0.4-14.4 µm optimized for<br />

measuring surface temperature, ocean colour, global vegetation, cloud characteristics, aerosol<br />

concentrations, temperature and moisture soundings, snow cover and ocean currents. The<br />

spatial resolution will be 250 m for band 1-2, 500 m for band 3-7 and 1000 m for band 8-36.<br />

EOS AM-1 MODIS <strong>data</strong> will be available in spring 2000.<br />

MERIS has radiation-sensitive arrays (CCD’s) that provides spatial sampling in the acrosstrack<br />

direction and the <strong>satellite</strong>’s motion provides scanning in the along- track direction. It<br />

will have 15 spectral bands in the range <strong>of</strong> 0.39-1.04 µm, that can be selected by ground<br />

command. The swath width will be 1150 km. The spatial resolution will be 300 m over<br />

coastal zones and land surfaces where communication capabilities exists. Otherwise, the<br />

resolution is reduced to 1200 m to reduce the amount <strong>of</strong> <strong>data</strong> that will be recorded onboard.<br />

ENVISAT MERIS <strong>data</strong> will be available in spring 2001.<br />

4


3. A review <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />

3.1 Introduction<br />

In conventional <strong>classification</strong> in remote sensing discrete pixels are used, i.e. the result is only<br />

one class per pixel. Much information about the memberships <strong>of</strong> the pixel to other classes is<br />

lost. This additional information can increase the accuracy <strong>of</strong> the <strong>classification</strong>.<br />

Mixed pixels or ‘mixels’ occur because the pixel size may not be fine enough to capture<br />

detail on the ground necessary for specific applications or where the ground properties, such<br />

as vegetation and soil types, vary continuously, as almost everywhere.<br />

Fuzziness suggests that a given pixel, owing to its spectral reflectance properties, may be<br />

placed into more than one informational/spectral class. The output <strong>of</strong> a <strong>s<strong>of</strong>t</strong> <strong>classification</strong> is a<br />

set <strong>of</strong> images (one per class) that express for each pixel the degree <strong>of</strong> membership in the class<br />

in question.<br />

S<strong>of</strong>t classifiers can be useful in delineating forest boundaries, shorelines and other<br />

continuous classes. They can also bring out objects that cover small areas, which with<br />

conventional classifiers otherwise would have disappeared.<br />

In training and testing in a <strong>classification</strong>, mixed pixels are usually avoided. But it may be<br />

difficult to acquire a training set <strong>of</strong> an appropriate size if only pure pixels are selected for<br />

training, since large homogenous regions <strong>of</strong> each class are needed in the image. The training<br />

statistics defined, may not be fully representative <strong>of</strong> the classes and so provide a poor base for<br />

the remainder <strong>of</strong> the analysis.<br />

By recognising that there are various degrees to which fuzziness may be incorporated in<br />

each stage, a continuum <strong>of</strong> <strong>classification</strong> fuzziness may be defined (Foody 1999). As<br />

fuzziness may be a characteristic feature <strong>of</strong> both the remotely sensed <strong>data</strong> and the ground<br />

<strong>data</strong>, the use <strong>of</strong> a fuzzy <strong>classification</strong> algorithm alone may be insufficient for the resolution <strong>of</strong><br />

the mixed-pixel problem.<br />

The continuum ranges from fully fuzzy, where fuzziness is accommodated in all three<br />

stages <strong>of</strong> the <strong>classification</strong> (the training-, allocation and testing stages), to completely crisp,<br />

which are the conventional <strong>classification</strong>.<br />

A modified maximum-likelihood classifier can accommodate fuzziness in any or all three<br />

stages <strong>of</strong> the analysis and neural networks can provide a <strong>classification</strong> at any point along the<br />

continuum <strong>of</strong> <strong>classification</strong> fuzziness.<br />

In training, the <strong>data</strong> is hard if the pixels are pure and vary in fuzziness for mixed pixels.<br />

Data on the fuzzy membership properties <strong>of</strong> training samples may be used to derive refined<br />

class descriptors for use in the class allocation stage. By refining the training statistics where<br />

some or even all <strong>of</strong> the pixels were mixed, the statistics that would have been derived if the<br />

pixels had in fact been pure can be obtained (Foody 1999).<br />

In the allocation stage for a <strong>s<strong>of</strong>t</strong> classifier a pixel can be allocated from a single class to all<br />

classes it provide membership in.<br />

This chapter will describe the theory and algorithms behind the existing <strong>s<strong>of</strong>t</strong> classifiers.<br />

To make the theory more understandable this chapter will also describe some applications that<br />

have been made.<br />

5


3.2 Fuzzy Set Theory<br />

Fuzzy sets are sets without sharp boundaries and they are applied to handle uncertainty in the<br />

process <strong>of</strong> <strong>classification</strong> (Palubinskas 1994). The result is <strong>of</strong>ten a more detailed and precise<br />

<strong>classification</strong>. Fuzziness can effectively extend the usefulness <strong>of</strong> map products developed<br />

from remote sensing imagery. The fuzzy set theory is particularly interesting as the analyst<br />

controls the degree <strong>of</strong> fuzziness (Foody 1996).<br />

It is the shape <strong>of</strong> the membership function that defines the ‘fuzziness’ <strong>of</strong> the phenomenon<br />

represented by the set. There are four membership functions that are usually used: sigmoidal<br />

(s-shaped), J-shaped, linear and user-defined (Eastman 1997). The sigmoidal membership<br />

function is the most commonly used in fuzzy set theory, see figure 2. It can be produced using<br />

a cosine function that requires the positions (along the X-axis) <strong>of</strong> 4 points governing the shape<br />

<strong>of</strong> the curve.<br />

The J-Shaped membership function is also quite common, although in most <strong>case</strong>s the<br />

sigmoidal function would be better. The linear is used extensively in electronic devices<br />

advertising fuzzy set logic and the user-defined is used when the relationships between the<br />

value and fuzzy membership does not follow any <strong>of</strong> the above three functions. The userdefined<br />

is the most applicable (Eastman 1997). The control points used in this function can be<br />

as many as necessary to define the fuzzy membership curve. The fuzzy membership between<br />

any two control points is linearly interpolated.<br />

Figure 2. The Sigmoidal Membership Function (<strong>based</strong> on Eastman 1997).<br />

The outputs for each pixel are the different membership values for each class. A high<br />

membership value indicates that the relevant land cover type covers a large area <strong>of</strong> the pixel<br />

(Bastin 1997). When using summation <strong>of</strong> memberships for a post-<strong>classification</strong> <strong>of</strong> fuzzy<br />

<strong>classification</strong> results, most <strong>of</strong> the increase <strong>of</strong> average <strong>classification</strong> accuracy is achieved after<br />

the first iteration (Palubinskas 1994).<br />

6


3.2.1 Fuzzy c-means (k-means)<br />

Fuzzy c-means is a clustering algorithm used for either unsupervised or supervised<br />

<strong>classification</strong>. The algorithm subdivides a <strong>data</strong> set into c-clusters or classes (Foody 1996). It<br />

begins by randomly assigning pixels to classes and iteratively moves the pixels to other<br />

classes with the aim <strong>of</strong> minimizing the generalised least-squared error,<br />

J m ( U,<br />

v)<br />

= ∑∑(<br />

u<br />

n<br />

c<br />

k = 1 i=<br />

1<br />

where U is a fuzzy c-partition <strong>of</strong> the <strong>data</strong> set Y containing n pixels, c is the number <strong>of</strong> classes,<br />

|| ||A is an inner product norm, v is a vector <strong>of</strong> cluster centres, vi is the centre <strong>of</strong> cluster i and<br />

m is a weighting component that lies within the range 1≤ m ≤ ∞ which determines the degree<br />

<strong>of</strong> fuzziness. The inner product norm is derived from,<br />

2<br />

7<br />

k<br />

)<br />

m<br />

y<br />

k<br />

− v<br />

T<br />

y − v = ( y − v ) A(<br />

y − v )<br />

k<br />

i<br />

A<br />

k<br />

−1<br />

A number <strong>of</strong> norms may be selected, for instance the Mahalanobis norm, A = C y , where<br />

Cy is the covariance matrix <strong>of</strong> the <strong>data</strong> set Y.<br />

To implement the fuzzy c-means algorithm, additional parameters are required to guide the<br />

partitioning process. These parameters are: selection <strong>of</strong> a distance measure and choosing a<br />

weighting exponent. The weighting exponent controls the ‘hardness’ or ‘fuzziness’ <strong>of</strong> the<br />

<strong>classification</strong>. The range <strong>of</strong> useful values for m is 1.5 < m < 3.0.<br />

The fuzzy c-means algorithm is particularly useful in circumstances where it is not<br />

reasonable to make assumptions about the statistical distributions <strong>of</strong> sample <strong>data</strong> (e.g. where<br />

training sets <strong>of</strong> pure pixels are small).<br />

For each pixel a fractional value is obtained for each class in the form <strong>of</strong> a real number<br />

between 0 and 1, and will generally sum up to 1.0 across all candidate classes.<br />

As output one also gets a residual error map showing the RMS error across the image.<br />

The values need to be adjusted (i.e. using the slope from linear regression) for area<br />

predictions to be made.<br />

3.2.2 Applications with Fuzzy c-means classifier<br />

A comparative <strong>study</strong> between traditional unsupervised <strong>classification</strong> and the fuzzy c-means<br />

algorithm was done by Manyara and Lein (1994). Landsat MSS subsets were conducted to<br />

detect environmental change. The fuzzy c-means were used since the boundary between forest<br />

and other categories <strong>of</strong> land cover may be difficult to delineate precisely. Four smaller <strong>study</strong><br />

sites depicting different forest environments were extracted from the image. The approach <strong>of</strong><br />

the fuzzy c-means proved that a pixel <strong>of</strong>ten belongs to more than one class. Of the total 2 500<br />

pixels per site, only 249 were pure in site I, 219 in site II, 29 in site III and 87 in site IV.<br />

Another comparison has been made by Bastin (1997) between the fuzzy c-means and the<br />

linear mixture model in the use for unmixing <strong>of</strong> coarse pixel signatures to identify four land<br />

cover classes. The Mahalanobis norm was used as a typically measure and after a little<br />

experimentation the weighting exponent, m, was set to 1.5. By this, dominant and minor land<br />

cover classes in each pixel was easily identified, without the <strong>classification</strong> being overly hard.<br />

The fuzzy c-means classifier performed best overall at locating and quantifying inclusions in<br />

mixed pixels.<br />

i<br />

k<br />

i<br />

i<br />

2<br />

A


3.3 S<strong>of</strong>tening maximum likelihood classifier (MCL)<br />

MCL operate by using band means and standard deviations from training <strong>data</strong> to model land<br />

cover classes as centroids in feature space, surrounded by probability contours. The<br />

probability density function assumes that the sample values for each class are normally<br />

distributed. The unclassified pixels are plotted in the same feature space and get a posteriori<br />

probability. Usually the pixels are then assigned to the class for which they have highest<br />

membership probability. But it is possible to <strong>s<strong>of</strong>t</strong>en the maximum likelihood <strong>classification</strong> by<br />

using the a posteriori membership probability values as indices <strong>of</strong> class membership (Bastin<br />

1997).<br />

The <strong>data</strong> must still satisfy the assumptions and requirements <strong>of</strong> the <strong>classification</strong> technique<br />

used, which is <strong>of</strong>ten unlikely with the probability-<strong>based</strong> classifiers.<br />

3.3.1 Applications with <strong>s<strong>of</strong>t</strong>ening MCL<br />

In the comparison made by Bastin (1997) described above, a shell script was used to combine<br />

each set <strong>of</strong> eight maximum likelihood <strong>classification</strong> output maps, and to produce a map<br />

showing the a posteriori probabilities for each class. The result was mostly extremely<br />

classified pixels, i.e. most <strong>of</strong> them were classified as pure pixels or as habitats being entirely<br />

absent. To avoid an incorrect evaluation <strong>of</strong> this method, the <strong>classification</strong> should be <strong>s<strong>of</strong>t</strong>ened,<br />

by broadening the spread <strong>of</strong> the normal distribution functions, which represent the land cover<br />

classes.<br />

3.4 Linear mixture modelling<br />

For linear unmixing <strong>of</strong> the spectral <strong>data</strong>, a ‘pure’ spectral signature needs to be obtained from<br />

the training <strong>data</strong>, for each <strong>of</strong> the defined land cover types (Bastin 1997). The model views<br />

mixed pixel signatures as being made up <strong>of</strong> a simple weighted linear sum <strong>of</strong> the component<br />

land cover signatures in that pixel. The weights are directly determined by the relative<br />

proportions <strong>of</strong> ground covered by each different land cover class by,<br />

x = Mf + e<br />

where x is the set <strong>of</strong> digital values measured by the sensor in each <strong>of</strong> n measurement<br />

wavelengths, f is a vector <strong>of</strong> ground cover proportions, for each <strong>of</strong> c land covers.<br />

M is a matrix <strong>of</strong> n x c coefficients, in which the columns are the vectors µ1…µc, which<br />

represent the pure spectral signatures given by the c cover classes in the absence <strong>of</strong> noise, e is<br />

a terments noise, which is minimized in some way in order to obtain a measure <strong>of</strong> the<br />

proportion vector f.<br />

The expected signature for a mixed pixel is the sum,<br />

f1µ 1 + f 2µ<br />

2 + ... + f c µ c = Mf<br />

where for class c, fc is the proportion <strong>of</strong> the pixel covered, and µc is the characteristic<br />

signature.<br />

The noise caused by the sensor and by natural variability within a scene is quantified as<br />

the vector <strong>of</strong> errors, e, which can be quantified as the sum <strong>of</strong> squares <strong>of</strong> its elements or as the<br />

variance <strong>of</strong> the training <strong>data</strong>.<br />

The linear mixture model assumes that each photon on the ground is interacting with only<br />

one land cover type before being reflected back to the sensor. In a complex and non-random<br />

landscape the linear mixture model is a reasonable approximation.<br />

The model is trained using the matrix <strong>of</strong> values M generated from ground truth<br />

8


information on ground cover proportions within selected pixels or from selected ‘purest’<br />

pixels for each component. The effectiveness <strong>of</strong> the model is dependent on the degree <strong>of</strong><br />

separation <strong>of</strong> the different signatures and the level <strong>of</strong> noise present in the scene.<br />

Rather than incorporating the noise and variability <strong>of</strong> the sample <strong>data</strong> into statistical or<br />

fuzzy limits around each cluster centriod, the linear mixture model uses the centroid vectors<br />

to define the spectral signatures <strong>of</strong> the pure land cover classes, and separates out scene/sensor<br />

noise into a separate error term.<br />

The number <strong>of</strong> distinguishable land cover types is strictly limited by the number <strong>of</strong> spectral<br />

bands in the <strong>data</strong>.<br />

3.4.1 Applications with linear mixture modelling<br />

The comparison, mentioned above in the fuzzy c-means algorithm application and in the<br />

<strong>s<strong>of</strong>t</strong>ening <strong>of</strong> MCL, showed that the linear mixture model assigned extreme values to most <strong>of</strong><br />

the pixels (Bastin 1997). One <strong>of</strong> the reasons for that is that there were pixels outside the range<br />

from zero to one in the <strong>classification</strong>, so called outliers. These outliers were reclassified to<br />

minimum or maximum values respectively. The result from the linear mixture model was<br />

fairly successful at picking up the general spatial patterns and gradations <strong>of</strong> the different cover<br />

classes, but it had problems with specific land cover signatures, because <strong>of</strong> poor spectral<br />

separability (Bastin 1997).<br />

3.5 Methods <strong>based</strong> on Bayesian probability theory<br />

The prime use for methods <strong>based</strong> on Bayesian probability theory is to determine the extent to<br />

which mixed pixels exist in the image, and their relative proportions. The methods can only<br />

be used when complete information is available or assumed (Eastman 1997). They are <strong>based</strong><br />

on the bayesian probability theory, which is the primary tool for the evaluation <strong>of</strong> the<br />

relationship between the indirect evidence and the decision set. The bayesian probability<br />

theory allows us to combine new evidence about a hypothesis along with prior knowledge to<br />

arrive at an estimate <strong>of</strong> the likelihood that the hypothesis is true. The basis for this is<br />

Bayes´theorem, which states that,<br />

p(<br />

h e)<br />

=<br />

where:<br />

p(h⏐e) = the probability <strong>of</strong> the hypothesis being true, given the evidence (posterior<br />

probability)<br />

p(e⏐h) = the probability <strong>of</strong> finding that evidence, given the hypothesis, being true<br />

p(h) = the probability <strong>of</strong> the hypothesis being true regardless <strong>of</strong> the evidence (prior<br />

probability)<br />

∑<br />

i<br />

p(<br />

e h)<br />

⋅ p(<br />

h)<br />

p(<br />

e h ) ⋅ p(<br />

h )<br />

The variance/covariance matrix derived from training site <strong>data</strong> is the matrix that allows one<br />

to assess the multivariate conditional probability p(e⏐h). This quantity is then modified by the<br />

prior probability <strong>of</strong> the hypothesis being true and then normalized by the sum <strong>of</strong> such<br />

considerations over all classes. This modification- and normalization step is important<br />

because it assumes that the only possible interpretation <strong>of</strong> a pixel is one <strong>of</strong> those classes for<br />

which training site <strong>data</strong> have been provided. Thus even weak support for a specific<br />

interpretation may appear to be strong if it is the strongest <strong>of</strong> the possible choices given. So<br />

even if the reflectance <strong>data</strong> for a certain class in a pixel is very weak it is treated as<br />

unequivocally belonging to that class (p=1.0) if no support exists for other classes. It therefore<br />

admits no ignorance and is a confident classifier.<br />

9<br />

i<br />

i


The posterior probability p(h⏐e) is the same quantity that the maximum-likelihood module<br />

evaluates to determine the most likely class.<br />

If a pixel has posterior probabilities <strong>of</strong> belonging to certain classes <strong>of</strong> 0.72 and 0.28<br />

respectively, this would be interpreted as evidence that the pixel contains 72 % <strong>of</strong> the first<br />

class and 28 % <strong>of</strong> the second. But this requires that the classes for which training site <strong>data</strong><br />

have been provided are the only one existing and that the conditional probability distribution<br />

p(e⏐h) do not overlap in the <strong>case</strong> <strong>of</strong> pure pixels. These requirements may in practice be very<br />

difficult to meet.<br />

3.5.1 Applications with Bayesian probability theory<br />

No reports on studies using Bayesian probability theory have been found.<br />

3.6 Methods <strong>based</strong> on Dempster-Shafer Theory<br />

The methods <strong>based</strong> on Dempster-Shafer theory are the most important <strong>of</strong> the <strong>s<strong>of</strong>t</strong> classifiers.<br />

It allows for the expression <strong>of</strong> ignorance in uncertainty management (Eastman 1997). The<br />

prime use <strong>of</strong> these methods is to check for the quality <strong>of</strong> one’s training site <strong>data</strong> and the<br />

possible presence <strong>of</strong> unknown classes.<br />

The basic assumptions <strong>of</strong> Dempster-Shafer theory are that ignorance exists in the body <strong>of</strong><br />

knowledge, and that belief for a hypothesis is not necessarily the complement <strong>of</strong> belief <strong>of</strong> its<br />

negation.<br />

The degree to which evidence provides concrete support for a hypothesis is known as<br />

belief, and the degree to which the evidence does not refute that hypothesis is known as<br />

plausibility. The difference between these two is the belief interval, which acts as a measure<br />

<strong>of</strong> uncertainty about a specific hypothesis.<br />

If the evidence supports one class to the degree <strong>of</strong> 0.3 and all other to 0.0, the Dempster-<br />

Shafer theory would assign a belief <strong>of</strong> 0.3 to that class and a plausibility <strong>of</strong> 1.0, yielding a<br />

belief interval <strong>of</strong> 0.7. Furthermore, it would assign a belief <strong>of</strong> 0.0 to all other classes and a<br />

plausibility <strong>of</strong> 0.7.<br />

Dempster-Shafer theory defines hypotheses in a hierarchical structure and will accept all<br />

possible combinations <strong>of</strong> hypotheses, see figure 3. The combinations <strong>of</strong> hypotheses<br />

acceptance is because it <strong>of</strong>ten happens that the evidence we have, supports some<br />

combinations <strong>of</strong> hypotheses without the ability to further distinguish the subsets. The basic<br />

classes are called singletons and the combinations non-singletons.<br />

[A,B,C]<br />

[A,B] [A,C] [B,C]<br />

[A] [B] [C]<br />

Figure 3. The Hierarchical Structure <strong>of</strong> the Subsets in the Whole Set [A,B,C] (<strong>based</strong> on<br />

Eastman 1997).<br />

10


When evidence provides some degree <strong>of</strong> commitment to one <strong>of</strong> these non-singleton<br />

classes and not to any <strong>of</strong> its constituents separately, this commitment is called Basic<br />

Probability Assignment (BPA). Thus, belief in a non-singleton class is the sum <strong>of</strong> BPA’s for<br />

that class and all sub-classes. Belief represents the total commitment to all members <strong>of</strong> a set<br />

combined.<br />

These non-singletons represents mixtures and might be used for more detailed examination <strong>of</strong><br />

sub-pixel <strong>classification</strong>.<br />

3.6.1 Applications with Dempster-Shafer theory<br />

In a <strong>study</strong> <strong>of</strong> fire scars in Indonesia by Fuller and Fulky (1998), Dempster-Shafer theory was<br />

used. That algorithm was used because each AVHRR pixel had a resolution <strong>of</strong> 1.1 km and<br />

could have contained several patches <strong>of</strong> fire scars. The only training sites used was for fire<br />

scars, all other classes were to be ‘other’ in the <strong>classification</strong>. As input some surface<br />

temperature layers and vegetation change detection layers were used. No rigorous ground<br />

validation could be done, due to lack <strong>of</strong> ground <strong>data</strong> at a scale meaningful to the AVHRR<br />

analysis. But compared to the other method used in the <strong>study</strong>, multiple threshold approach,<br />

the Dempster-Shafer method managed to extract more fire scars and came closer to the result<br />

done by the the Center for Remote Sensing, Imaging and Processing (CRISP) <strong>of</strong> the National<br />

University <strong>of</strong> Singapore. The CRISP analysis relied on visual interpretation <strong>of</strong> 766 SPOT<br />

quicklook images mapped to 100 m spatial resolution.<br />

3.7 Neural Network<br />

Neural networks learn by example (Kanellopoulos 1997). From the training set the network<br />

can learn the values <strong>of</strong> the samples internal parameters. Usually neural networks deal with<br />

large amounts <strong>of</strong> training <strong>data</strong> (i.e. thousands <strong>of</strong> samples) whereas statistical methods use<br />

much smaller training sets. The goal is the representation <strong>of</strong> complicated phenomena. The<br />

neural network does not make any explicit a priori assumptions about the <strong>data</strong>.<br />

3.7.1 The Multilayer Perceptron Classifier using feed-forward and back-propagation<br />

The multilayer perceptron classifier (MLP) are the most widely used neural classifier today.<br />

These networks are general-purpose, flexible, non-linear models consisting <strong>of</strong> a number <strong>of</strong><br />

units organised into multiple layers (Kanellopoulos 1997). Varying <strong>of</strong> the number <strong>of</strong> layers<br />

and the number <strong>of</strong> units in each layer can change the complexity <strong>of</strong> the MLP. These networks<br />

are valuable tools in problems when one has little or no knowledge about the form <strong>of</strong> the<br />

relationship between input vectors and their corresponding outputs. The principle for the<br />

multi-layer perceptron neural network is that processing elements or nodes divided into layers<br />

perform calculations until an output value is computed at each <strong>of</strong> the output nodes (Foody<br />

1996). Data to the input layer are the <strong>satellite</strong> channel values used in the <strong>classification</strong><br />

procedure. Then one or more hidden layers perform the calculations and send the result to the<br />

output layer, which has one node for each class. Every node in a layer is connected to every<br />

node in the layer above and below, see figure 4. The connections carry weights, which<br />

encapsulate the behaviour <strong>of</strong> the network and are adjusted during training.<br />

11


Level<br />

k<br />

j<br />

i<br />

Connection<br />

weights<br />

Output Classes<br />

1 2<br />

Input Pattern Feature Values<br />

Figure 4. The Architecture <strong>of</strong> Multi-Layer Perceptron (<strong>based</strong> on Kanellopoulos 1997).<br />

For the hidden layers the input to each node is the sum <strong>of</strong> the scalar products <strong>of</strong> the<br />

incoming vector components with their respective weights,<br />

input<br />

∑<br />

= i<br />

where wji is the weight connecting node i to node j and outi is the output from node i.<br />

The output <strong>of</strong> a node j is,<br />

The function f denotes the activation function <strong>of</strong> each node. The most frequently used is the<br />

sigmoid activation function,<br />

where x =inputj. This ensures that the node acts like a thresholding device.<br />

The network learns using the back-propagation algorithm iteratively. This algorithm<br />

minimizes the mean square error between the network's output and the desired output,<br />

where P are all input patterns and dk is the desired output.<br />

j<br />

12<br />

w<br />

ji<br />

out<br />

out j = f ( input j )<br />

1<br />

f ( x)<br />

=<br />

1+<br />

exp( −x)<br />

1<br />

E = ∑∑( d k − out k )<br />

2P<br />

P k<br />

i<br />

2<br />

Output layer<br />

Hidden layer<br />

Input layer


It compares the calculated output with the desired one, and adjust the weights attached to<br />

the connections, until the difference between the outputs is reduced to an acceptable level and<br />

the set <strong>of</strong> weights is stable.<br />

The weights are from the beginning small and random and then adjusted by the<br />

generalised delta rule,<br />

where wkj (t+1) and wkj (t) are the weights connecting nodes k and j at iteration (t+1) and t<br />

respectively, η is a learning rate parameter.<br />

For nodes in the output layer,<br />

and for nodes in the hidden layers,<br />

wkj ( t + 1)<br />

= wkj<br />

( t)<br />

+ η(<br />

δ koutk<br />

)<br />

δ<br />

k<br />

= ( d k − outk<br />

) f ′ ( inputk<br />

)<br />

To measure the generalisation ability it is common to have a set <strong>of</strong> <strong>data</strong> to train the network<br />

and a separate set to assess the performance. Once the neural network has been trained, the<br />

weights are saved to be used in the <strong>classification</strong> phase.<br />

The activation level <strong>of</strong> an output unit is positively related to the strength <strong>of</strong> membership to<br />

the class associated with the unit and lies on a 0-1 scale. These levels are significantly<br />

correlated with the sub-pixel land cover composition and may be mapped as fraction images.<br />

The classifier encounters lot <strong>of</strong> problems to separate the classes when the classes have lot<br />

<strong>of</strong> similarities. A-priori knowledge <strong>of</strong> the relative class distribution could be used to apply<br />

non-random weights to the network inputs to further distinguish similar classes.<br />

Another disadvantage is that the back-propagation method is extremely slow. About<br />

thousand iterations need to be done before the network converges to a solution. Sometimes no<br />

solution is found (Augusteijn and Warrender 1998).<br />

3.7.2 Applications with Multilayer Perceptron Classifier<br />

Several applications have been performed with the multilayer perceptron classifier. Howald<br />

classified Landsat TM <strong>data</strong> into seven land cover classes using a three layer network, with a<br />

slightly better overall <strong>classification</strong> accuracy than a maximum likelihood classifier (1989).<br />

Kanellopoulos et al. used a four-layer neural network for classifying multi-temporal SPOT<br />

multi-spectral imagery into twenty land cover classes (1991). Owing to the complexity <strong>of</strong> the<br />

<strong>classification</strong>, a relatively large neural network architecture was required and therefore the<br />

computation time was quite long. Gualtieri and Withers used clustering <strong>of</strong> multispectral <strong>data</strong><br />

(1988). Ryan et al. (1991) reported the use <strong>of</strong> a neural network to categorise small blocks <strong>of</strong><br />

pixels to extract shoreline features from images.<br />

Foody and Boyd (1999) used a neural network to derive fuzzy <strong>classification</strong>s <strong>of</strong> land cover<br />

along a transect crossing the transition from moist semi-deciduous forest to savannah in West<br />

Africa in February and December 1990. They used NOAA AVHRR for the <strong>study</strong>. The<br />

training <strong>data</strong> was acquired from core regions <strong>of</strong> each class. The AVHRR provides five<br />

spectral channels, so the network had five input units. The architecture also consisted <strong>of</strong> five<br />

hidden units in a single layer and four output units and a logistic sigmoid transfer function<br />

was used. The initial network weights were set randomly between the limits <strong>of</strong> ± 0.5 and the<br />

learning rate and momentum were set to 0.1 and 0.9 respectively. When 1000 iterations <strong>of</strong><br />

13<br />

∑<br />

δ f ′ (<br />

input ) δ w<br />

j = k k kj


stochastic backpropagation learning algorithm had been performed, the RMS error was 0.096<br />

and 0.117 for the February and December <strong>data</strong> sets respectively. The method was able to<br />

represent both gradual and sharp changes in land cover, as those associated with the forestsavannah<br />

and savannah-water boundaries respectively. It also enabled important properties <strong>of</strong><br />

the boundaries to be inferred, such as that the width <strong>of</strong> the forest-savannah transitional zone<br />

remained relatively constant, but had migrated significantly, between the two time periods.<br />

3.7.2 Neural Network using Cascade-correlation<br />

Improvements have been made on back-propagation, since it is so slow. One <strong>of</strong> these<br />

improved architectures is the cascade-correlation (Augusteijn and Warrender 1998). Cascadecorrelation<br />

is a feed-forward neural network. The network builds its internal structure<br />

incrementally, during training. The initial network consists <strong>of</strong> only two layers: an input layer<br />

and an output layer, which are completely connected. These connections are trained until no<br />

more significant changes occur between iterations. If the total training error still is too high, a<br />

hidden node will be allocated and trained to further reduce this error. This goes on until the<br />

termination criterion is reached or until a maximum number <strong>of</strong> hidden nodes are reached and<br />

the conclusion is made that the network cannot learn the problem.<br />

3.7.3 Applications with Cascade-correlation<br />

A <strong>study</strong> made <strong>of</strong> Augusteijn and Warrender (1998) was conducted to investigate the ability <strong>of</strong><br />

a cascade-correlation neural network to delineate upland and forested wetland areas and to<br />

distinguish between different levels <strong>of</strong> wetness in a forested wetland. They used NASA’s<br />

Airborne Terrestrial Applications Sensor (ATLAS) multispectral <strong>data</strong> and Airborne Imaging<br />

Radar Synthetic Aperture Radar (AIRSAR) <strong>data</strong>. The input values used for training and<br />

testing <strong>of</strong> the neural network classifier were not simply digital number (DN) values indicating<br />

the reflectance <strong>of</strong> each pixel in a given band, but averages <strong>of</strong> DN values calculated over 3<br />

pixel by 3 pixel neighbourhoods inside the training and test polygons. The same sample size<br />

had to be used in each category for the network to learn all categories equally well. First a five<br />

class categorization was tried. But three <strong>of</strong> the five classes were too similar to be<br />

distinguishable, so they were merged together. Several networks were trained using different<br />

band combinations. Every network were always trained five times on the same <strong>data</strong> set, each<br />

time using a different selection <strong>of</strong> random initial values for the connection strengths, to reduce<br />

the influence <strong>of</strong> a specific set <strong>of</strong> initial values on network performance. Test sample<br />

agreement scored well over 80 per cent in most <strong>case</strong>s.<br />

3.7.4 Fuzzy ARTMAP<br />

ARTMAP achieves a synthesis <strong>of</strong> fuzzy logic and ART networks (Mannan et al. 1998). ART<br />

stands for Adaptive Resonance Theory. The architecture <strong>of</strong> fuzzy ARTMAP consists <strong>of</strong> four<br />

layers <strong>of</strong> neurons: the input-, category-, mapfield-, and output layer. The values <strong>of</strong> the spectral<br />

bands and their complement, <strong>of</strong> dimension n, are the input to the first layer, which consists <strong>of</strong><br />

2n neurons. The category layer starts with one neuron, but grows in number as the learning<br />

proceeds. The output and mapfield layers consist <strong>of</strong> as many neurons as there are classes.<br />

Two vigilance parameters ρ1 and ρ2 control the operation during learning and operational<br />

phases <strong>of</strong> the network.<br />

14


The learning phase is an iteratively process where for each input a category choice is<br />

calculated according to,<br />

A ∧W<br />

S =<br />

α + W<br />

where A is the input feature vector and W1 is the weight vector between the input layer and a<br />

node in the category layer.<br />

For the node which has the largest value <strong>of</strong> S the match ratio at mapfield is calculated by,<br />

where B is the output class vector and W2 is the weight vector between a chosen node <strong>of</strong><br />

category layer and the mapfield layer.<br />

If the Rm value is larger than ρ2 the weights are changed and a new input value is managed.<br />

The learning process <strong>of</strong> mapfield- and category layer weights is called resonance, and the<br />

weights are calculated by,<br />

( new)<br />

1<br />

If not, ρ1 = Rc for that input, where Rc is calculated by,<br />

1<br />

R m<br />

For the nodes which has Rc ≥ ρ1 it is checked if any has Rm ≥ ρ2, if so is the <strong>case</strong> the<br />

weights are changed and a new input is managed, otherwise a new nodes committed, the<br />

weights are changed and a new input managed. This goes on until all training samples are<br />

exhausted and the category layer nodes stop growing or the number <strong>of</strong> iterations exceeds T, a<br />

chosen positive constant.<br />

After this process a score S is calculated for each <strong>of</strong> the committed nodes. The node in the<br />

output layer, which corresponds to the node with the largest value <strong>of</strong> S, indicates the category<br />

<strong>of</strong> the input pixel.<br />

α can be chosen around 0.01, β can be set to 1.0 in the beginning <strong>of</strong> the learning phase and<br />

to a smaller value later. ρ1 and ρ2 are set to 1.0 for the most accurate results.<br />

There is an advantage compared to maximum-likelihood and the multilayer perceptron<br />

classifier (MLP) <strong>of</strong> about 5 % in the overall accuracy (Mannan et al. 1998). The method is<br />

stable, easy to use, with a smaller number <strong>of</strong> parameters to manage and faster than MLP.<br />

15<br />

1<br />

1<br />

B ∧W<br />

=<br />

B<br />

( old )<br />

1<br />

∗<br />

2<br />

W = β ( A ∧W<br />

) + ( 1−<br />

β ) W<br />

W = β ( A ∧W<br />

) + ( 1−<br />

β ) W<br />

( new)<br />

2<br />

2<br />

R c<br />

( old )<br />

2<br />

A ∧W1<br />

=<br />

A<br />

1<br />

2<br />

( old )<br />

1<br />

( old )<br />

2


3.7.5 Applications with Fuzzy ARTMAP<br />

Six multispectral images acquired by Linear Imaging Self-scanning Sensor (LISS-II) camera<br />

<strong>of</strong> Indian Remote Sensing Satellite (IRS-1B) have been used in an experiment in using fuzzy<br />

ARTMAP in <strong>classification</strong> by Mannan et al. (1998). The training samples were selected by<br />

visual interpretation <strong>of</strong> the scenes. In order to preserve the samples relative values, the grey<br />

levels were normalized. 40 per cent <strong>of</strong> the samples were used in the training phase and the rest<br />

were applied in the operational phase to assess the accuracy. The four spectral band pixel grey<br />

values and their complements were input to the network. In the output binary vector, the bit,<br />

which belongs to the input’s class, was set to 1 and the rest to 0:s.<br />

All weights were initially set to 1, but decreased as the learning progressed. The result was<br />

about 5 per cent better in the overall accuracy compared to multilayer perceptron- and<br />

maximum likelihood classifier. In this experiment 7483 pixels were correctly classified and<br />

397 misclassified.<br />

16


4. Data and methods<br />

4.1 Introduction<br />

This chapter describes the <strong>study</strong> area and the <strong>data</strong> that has been used in the <strong>classification</strong>.<br />

Two different <strong>s<strong>of</strong>t</strong>wares have been used to manage the problem and the method to do this, is<br />

described.<br />

4.2 Data<br />

4.2.1 Satellite <strong>data</strong><br />

For the <strong>case</strong> area, only one Indian IRS-1C image is available and that one is very cloudy.<br />

Since IRS-1C Wifs only have two spectral bands, at least two images are needed to get a<br />

reliable <strong>classification</strong>. So the choice has fallen on the Russian RESURS-O1 #3 with MSU-SK<br />

sensored images. Several RESURS images are available for the <strong>study</strong> area, but only three are<br />

more or less cloud-free. These are from July and August 1995 and September 1996. These<br />

RESURS images have five spectral bands, but only four can be acquired in parallell. Of these<br />

four spectral bands, only three are acceptably noise free to be used, these are the red band and<br />

the two near infrared bands, see figure 5. Only seven bands are allowed in the <strong>s<strong>of</strong>t</strong>ware, so the<br />

image from August 1995 is excluded.<br />

The <strong>study</strong> area constituted only a small part <strong>of</strong> each image and was therefore extracted to<br />

reduce the amount <strong>of</strong> <strong>data</strong>. The area was 22 km × 44 km.<br />

17<br />

Figure 5. RESURS multitemporal colour<br />

composite <strong>of</strong> the first near-infrared band,<br />

red band and the second near infrared<br />

band from July 1995 (RGB) over the <strong>study</strong><br />

area.


4.2.2 Reference <strong>data</strong><br />

As reference <strong>data</strong> for training and<br />

evaluation the existing land cover <strong>data</strong><br />

from the pilot production <strong>of</strong> the Swedish<br />

CORINE land cover over the Gripen area<br />

was used, see figure 6.<br />

CORINE stands for ‘Co-ordinating<br />

Information on the Environment’ and was<br />

a EU programme beginning 1985. This<br />

land cover is mainly <strong>based</strong> on information<br />

from digital topographical maps<br />

interpretation and <strong>classification</strong> <strong>of</strong> Landsat<br />

TM images with 30 m resolution,<br />

resampled to a pixelsize <strong>of</strong> 25 m.<br />

In most <strong>of</strong> Europe the CORINE land<br />

cover contains 44 classes, but Sweden has<br />

added 11 classes to better fulfil the needs<br />

(SSC 1998).<br />

The main classes are produced by: Figure 6. The reference image,<br />

Swedish CORINE land cover.<br />

♦ The wetland classes are mapped using masks from the digital topographic map and by<br />

classifying <strong>satellite</strong> <strong>data</strong>. The map provides the spatial limits and the <strong>satellite</strong> <strong>data</strong> is used<br />

to update the masks for wet mires and dry mires and for division <strong>of</strong> those classes. The<br />

mires near a lake or sea is by a GIS-operation named inland- or coastal marshes.<br />

♦ The water bodies are classified by thresholding in TM band 5, where that class has low<br />

digital values. Then the water is classified and subtracted from the water mask in the<br />

topographic map. The reason for this is that floating vegetation or lakes overgrown with<br />

reeds are obtained, and can be candidates for the inland marshes.<br />

♦ Forest is classified using codes from National Forest Survey field <strong>data</strong>. For given<br />

coordinates, these codes say what kind <strong>of</strong> forest that grows there. For the same<br />

coordinates, spectral signatures are extracted. By this the signatures for different kinds <strong>of</strong><br />

forest are known and the image can be classified under the forest mask. Filtered signatures<br />

are used.<br />

♦ Clear-cut forest is extracted from an image showing change, which is produced using<br />

histogram matching between the same channel in each scene. Filtered signatures are used<br />

for <strong>classification</strong> (nearest-neighbour).<br />

♦ Urban boundaries are interpreted on screen using the Statistics Sweden urban <strong>data</strong>base<br />

and <strong>satellite</strong> images as background. Buildings within 200 m are included.<br />

♦ Agricultural areas are taken from the digital topographical map, when it exists, as in my<br />

<strong>study</strong> area.<br />

The forest mask also contains the roads. These roads are mostly classified as forest<br />

regeneration or clear-cut forest. Narrow and long areas <strong>of</strong> these classes which coincides with<br />

18


oads narrower than 7 meters are alloted to the surrounding class. Then all areas in the image<br />

that were smaller than 2 pixels were removed.<br />

The received image was again resampled, now to 12.5 m.<br />

4.3 Study area<br />

4.3.1 General description <strong>of</strong> the Baltic Sea region landscape<br />

The Baltic Sea region is the area (i.e. countries) surrounding the Baltic Sea. From the<br />

vegetation’s point <strong>of</strong> view the countries are very young, in particularly Sweden, because <strong>of</strong><br />

the glaciation. The immigration <strong>of</strong> plants has occurred in several steps and from different<br />

directions (Bråkenhielm et al. 1998). The spreading <strong>of</strong> the plants is strongly dependent <strong>of</strong> the<br />

climate and their ability to compete with other species. The human activities, like agricultural<br />

and forestry have also affected the vegetation. Where a species chooses to grow is for<br />

example a matter <strong>of</strong> the bedrock, i.e. the content <strong>of</strong> nutrition in the mineral and rock and their<br />

ability to disintegrate and contain water. The disintegration give rise to the foundation <strong>of</strong> fine<br />

granular soils, which in turn helps keeping water, which in turn makes the chemical<br />

disintegration easier and results in more nutrition. The bedrock has developed from magmatic,<br />

sedimentary and metamorphic processes during billions <strong>of</strong> years. Different phases have<br />

contributed to bedrock <strong>of</strong> varying composition. This, along with agricultural fields with<br />

groups <strong>of</strong> trees and the variations in the terrain, makes the landscape patchy. How patchy a<br />

person or animal experience the landscape is due to the person’s references or the animals<br />

field <strong>of</strong> view or habitat. For example, doesn’t a hawk think the same area is patchy as do the<br />

mouse.<br />

This patchy landscape makes it difficult to generate a good and reliable hard <strong>classification</strong>,<br />

from medium resolution <strong>satellite</strong> images, since they consists <strong>of</strong> pixels with sizes between 150-<br />

250 m. It is not <strong>of</strong>ten such pixels contain just one class. Therefore, the main task for this<br />

project is to analyze if <strong>s<strong>of</strong>t</strong> classifiers can better show the real situation for such <strong>satellite</strong><br />

images.<br />

The <strong>study</strong> area is very suitable for this task, since it contains all kinds <strong>of</strong> land covers, from<br />

large agricultural fields with groves <strong>of</strong> trees, to large forests with patches <strong>of</strong> agricultural<br />

fields. In figure 7, it can be seen that there are several classes in a 150 m- pixel and therefore a<br />

hard classifier is not suitable.<br />

The <strong>study</strong> area, called Gripen, is situated between lake Hjälmaren and the southern part <strong>of</strong><br />

the Baltic Sea bay – Bråviken, in the southern part <strong>of</strong> Sweden. The area contains two big<br />

cities – Örebro and Norrköping. Around these cities, the landscape is very flat with large areas<br />

<strong>of</strong> arable land. The forested area between these cities is very patchy with many smaller lakes<br />

and hills. The forest contains both deciduous and coniferous forest and a mixture <strong>of</strong> both with<br />

no sharp boundaries.<br />

There are also different kinds <strong>of</strong> wetlands, which are <strong>of</strong> special interest in the BALANS<br />

project, since they act as filters <strong>of</strong> nutrients before the water runs into the Baltic Sea.<br />

19


Figure 7. A part <strong>of</strong> the <strong>study</strong> area in Gripen showing the reference image with 12.5 m pixel<br />

size and a 150 m grid to elicit the fuzziness in the medium resolution <strong>satellite</strong> image pixels<br />

(adapted from http://www.grida.no/baltic/htmls/maps.htm).<br />

4.3.2 Quantitative characterisation<br />

The area distribution <strong>of</strong> each class in the reference image are: water 5.19%, urban 1.82%,<br />

agricultural areas 20.08%, marsh 0.71%, deciduous forest 11.79%, coniferous forest 36.80%,<br />

mire 2.42%, clear-cut forest 7.35%, rock 0.82%, barren land 0.63%, peat 0.44% and grass<br />

3.51%. This result in 13% unknown classes when using eight classes and 9% unknown<br />

classes when using twelve classes.<br />

4.4 Tools<br />

4.4.1 IDRISI<br />

IDRISI is a geographic information and image processing <strong>s<strong>of</strong>t</strong>ware system developed by the<br />

Graduate School <strong>of</strong> Geography at Clark University, Worcester, U.S.A. It covers the full<br />

spectrum <strong>of</strong> GIS and remote sensing needs. The <strong>s<strong>of</strong>t</strong>ware has over 150 modules that provide<br />

facilities for the input, display and analysis <strong>of</strong> geographic <strong>data</strong>. For this project the fuzzy<br />

signature development and <strong>s<strong>of</strong>t</strong> classifier modules has been especially important. But at least<br />

this version, version 2.0 for Windows, has its limitations, which force the users to find other<br />

20


solutions to their problems, and sometimes import <strong>data</strong> from other systems, for example<br />

already radiometric- and geometric corrected images.<br />

The module <strong>based</strong> on the logic <strong>of</strong> Fuzzy Sets in IDRISI is FUZCLASS. In that module fuzzy<br />

set membership is determined from the distance <strong>of</strong> pixels from signature means. To use this<br />

module two parameters has to be set. The first is the z-score distance where fuzzy<br />

membership becomes zero. If a pixel has the same location as the class mean the membership<br />

grade is 1.0, and away from this point the grade decreases until it reaches zero. The second<br />

parameter is whether or not the membership values should be normalized. Normalization<br />

makes the assumption that the classes are the only one existing, and thus the membership<br />

values for all classes for a single pixel must sum to 1.0.<br />

The module using the Dempster-Shafer theory in IDRISI is BELCLASS. In normal use,<br />

Dempster-Shafer theory requires the classes under consideration to be mutually exclusive and<br />

exhaustive. But in BELCLASS the pixel may belong to some unknown class, for which a<br />

training site has not been provided. An additional category is added to every analysis called<br />

[other]. The result is consistent with Dempster-Shafer theory, but recognizes the possibility<br />

that there may be classes present about which we have no knowledge. The uncertainty is<br />

almost identical to that <strong>of</strong> Dempster-Shafer ignorance.<br />

In BELCLASS you have two choices <strong>of</strong> output: beliefs or plausibilities. In either <strong>case</strong>, one<br />

image <strong>of</strong> belief or plausibility is produced for each class. A <strong>classification</strong> uncertainty image is<br />

also produced.<br />

4.4.2 ERDAS IMAGINE<br />

ERDAS Imagine is a complete GIS and image processing <strong>s<strong>of</strong>t</strong>ware produced by the<br />

Engineering department <strong>of</strong> ERDAS, Inc., Atlanta, U.S.A. It is easy to use and has modules for<br />

most problems. The limitation in this <strong>case</strong> is that it only has one fuzzy module. For this<br />

project it has been used for preparation <strong>of</strong> the images and extraction <strong>of</strong> pixel information in<br />

the resulting images.<br />

21


4.5 Methodology<br />

The outline <strong>of</strong> the used method is shown in figure 8. A scheme showing the main action<br />

steps is included as appendix 1. The produced macro files are listed by name in a scheme in<br />

appendix 2.<br />

4.5.1 Outline<br />

Figure 8. The used methodology.<br />

4.5.2 Initial <strong>data</strong> preprocessing<br />

4.5.2.1 Reference <strong>data</strong><br />

Preprocessing<br />

reference image<br />

Binarisation,<br />

Thematic aggregation<br />

Spatial aggregation<br />

Sampling <strong>of</strong> training<br />

sites<br />

Preparation <strong>of</strong> training statistics<br />

Running the modules<br />

Evaluation and<br />

comparison <strong>of</strong> <strong>s<strong>of</strong>t</strong><br />

classifiers<br />

4.5.2.1.1 Binarisation<br />

The preprocessing <strong>of</strong> the reference image began with extraction <strong>of</strong> clouds that were in the<br />

RESURS images. This was done by using the mask made from the <strong>satellite</strong> images. To m<br />

make one image layer for each class a re<strong>classification</strong> was made. The result was a binary<br />

image for each class.<br />

22<br />

Preprocessing<br />

<strong>satellite</strong> images<br />

Radiometric correction<br />

Geometric correction


4.5.2.1.2 Thematic aggregation<br />

The reference image was reclassified to contain eight or twelve classes. These classes were<br />

water, urban, agricultural field (agri), marshes (marsh), deciduous forest (decid), coniferous<br />

forest (conif), wet/other mires (mire) and clear-cut forest (clear). The urban class consisted<br />

<strong>of</strong> the classes coarse- and dense city structures and smaller communities from the Swedish<br />

CORINE land cover. The coniferous and deciduous class consisted <strong>of</strong> conifers/decids with<br />

different age and height. When 12 classes were used, rock, barren land (barr), peat and<br />

grass were added.<br />

The choices were made <strong>based</strong> on the more detailed classes that the BALANS project<br />

wants to extract.<br />

The BALANS project has envisaged at least five broad classes.<br />

These are,<br />

• Artificial surfaces<br />

• Agricultural areas<br />

• Forests and semi-natural areas<br />

• Wetlands<br />

• Water bodies<br />

More detailed classes envisaged if possible is urban fabric, arable land, inland wetlands,<br />

coniferous- and deciduous forest. Therefore, a <strong>classification</strong> <strong>of</strong> these more detailed classes<br />

was tried out to.<br />

4.5.2.1.3 Spatial aggregation<br />

The reference image contained pixels with a size <strong>of</strong> 12.5 meter. To be able to compare the<br />

reference image with the RESURS images, which have a pixel size <strong>of</strong> 150 meter, an<br />

aggregation was made in IDRISI. In the process an average <strong>of</strong> twelve 12.5 m pixels in<br />

both the X- and Y-direction was calculated and was output as a 150 m pixel. That made<br />

the border pixels <strong>of</strong> each class patch fuzzy, i.e. they didn’t contain 100% <strong>of</strong> that class any<br />

more. One such layer were made for each class and these were seen as the ‘true’ fuzzy<br />

classes, since now the percent <strong>of</strong> fuzziness in each pixel was known.<br />

4.5.2.2 Satellite <strong>data</strong><br />

4.5.2.2.1 Radiometric correction<br />

Since RESURS images from two different dates were used they had different illumination.<br />

Therefore the image which visually seemed to be the best was set as reference in this <strong>case</strong>.<br />

Each band in the other image was histogram-matched against respectively band in the<br />

’reference’ image.<br />

4.5.2.2.2 Geometric correction<br />

The RESURS images didn’t have any known projection and didn’t lay above each other. In<br />

one image there was one band that didn’t lay above the other bands either. So the images<br />

were geometrically corrected image to image by taking ground control points (GCPs). The<br />

RMS error were about 25 m as maximum, that is about one sixth <strong>of</strong> a pixel. The images<br />

were then resampled with cubic convolution and then corrected against the reference image.<br />

The reference image was projected with Transverse Mercator and had datum and spheroid<br />

WGS 84, so after the correction against the reference image, the RESURS images received<br />

the same projection.<br />

A cloud mask was created by taken training sites in the clouds and other ‘close’ classes.<br />

A supervised <strong>classification</strong> with maximum-likelihood was performed and the output classes<br />

were tested to see which were actually clouds. These clouds were marked in an ‘area-<strong>of</strong>interest’-layer<br />

and cut out in all images.<br />

23


4.5.3 Choice <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />

The <strong>s<strong>of</strong>t</strong>ware available for <strong>s<strong>of</strong>t</strong> <strong>classification</strong> was IDRISI. This <strong>s<strong>of</strong>t</strong>ware included <strong>s<strong>of</strong>t</strong><br />

classifiers using Bayesian probability theory and Dempster-Shafer theory reviewed earlier. It<br />

also included one method where the distance from a class mean were set where the fuzzy<br />

membership <strong>of</strong> that class should be zero.<br />

4.5.4 Application <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />

4.5.4.1 Sampling <strong>of</strong> ’training <strong>data</strong>’ from reference <strong>data</strong><br />

Different training <strong>data</strong> samplings were examined.<br />

♦ The first sampling was to take random sample points in the aggregated masks, as training<br />

sites. In this sampling as well as in sampling two and three, different points were sampled<br />

for each class.<br />

♦ The second sampling was to make majority masks from the aggregated ones, were only<br />

the pixels with class values <strong>of</strong> 75-100 % (90-100% for water) were included, to receive<br />

purer sample points to see if there were any improvement.<br />

♦ The third sampling was to, first (in ERDAS Imagine) manually take training sites within<br />

the wetland mask to see where the wetland had a separable signature from the other<br />

classes. Then a mask from these wetland areas were made under the whole wetland layer<br />

and used as input to the aggregation in IDRISI. The sample points were taken in the<br />

majority masks.<br />

♦ The fourth sampling was to take random sample points in the whole image and divide<br />

them into eight equally quantified classes. In this <strong>case</strong> the points for a specific class could<br />

be sampled where that class didn’t have any pixel values. Most <strong>of</strong> the time the training<br />

samples for a specific class did not contain a majority <strong>of</strong> that class. All training samples<br />

were quite equal.<br />

4.5.4.2 Preparation <strong>of</strong> training statistics<br />

Each class’ training sites were seen as one training site group. After the sampling step, every<br />

training site group was compared to the aggregated masks to see how much <strong>of</strong> each class<br />

there was in every training site group. An average value was calculated for each class for each<br />

training site group, e.g. the output is an average value <strong>of</strong> content <strong>of</strong> water in the urban training<br />

site group and the content <strong>of</strong> water in the agricutural training site group etc. These average<br />

values were imported to the fuzzy partition matrix and linked to the image with the training<br />

site groups. The fuzzy partition matrix indicates the membership grades <strong>of</strong> each training site<br />

group in each class (Eastman 1997). After the linking to the training site groups image, one<br />

image for each class is created showing the membership grade in each training site group. The<br />

fuzzy signatures are created by giving each training site group a weight proportional to its<br />

membership grade when determining the mean, variance and covariance <strong>of</strong> each <strong>satellite</strong><br />

spectral band for each class.<br />

4.5.4.3 Running the modules<br />

After the creation <strong>of</strong> fuzzy signatures the different modules were tested. When the proportion<br />

<strong>of</strong> each class in the reference image is known, a prior probability can be input as a weight and<br />

improve the result.<br />

24


4.5.4 Evaluation and comparison <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />

4.5.4.1 Introduction to thematic accuracy assessment <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers<br />

The ground <strong>data</strong> can rarely be assumed to be error free. So testing a <strong>classification</strong> is an<br />

evaluation <strong>of</strong> the level <strong>of</strong> agreement or correspondence between two sets <strong>of</strong> class allocations,<br />

both <strong>of</strong> which have their own error characteristics, rather than quantifying the accuracy.<br />

In hard <strong>classification</strong>, accuracy assessment is generally made with the use <strong>of</strong> an error<br />

matrix. Error matrices compare, on a class-by-class basis, the relationship between known<br />

reference <strong>data</strong> and the result <strong>of</strong> a <strong>classification</strong>. Such matrices are square, with as many rows<br />

and columns as there are classes to be assessed. Several characteristics are expressed by the<br />

error matrix, for example the error <strong>of</strong> omission (exclusion) and error <strong>of</strong> commission<br />

(inclusion). Then the producer’s accuracy, user’s accuracy and overall accuracy can be<br />

calculated. The producer’s- and user’s accuracy tells how much <strong>of</strong>, for example the<br />

coniferous class, that is correctly classified and how much <strong>of</strong> the classified category<br />

coniferous that is truly that in reality, respectively.<br />

Another measure to extract from the error matrix is the kappa statistic, which is a measure<br />

<strong>of</strong> the difference between the observed accuracy and a chance agreement between the<br />

reference <strong>data</strong> and a random classifier.<br />

The problem is that the error matrix is only appropriate for use with hard <strong>classification</strong><br />

(Foody 1996). In <strong>s<strong>of</strong>t</strong> <strong>classification</strong> you receive one matrix for each class, i.e. each pixel<br />

would be compared. But each pixel is just allowed to consist <strong>of</strong> one class, consequently to be<br />

correct or incorrect.<br />

A number <strong>of</strong> approaches have been suggested to assess the accuracy <strong>of</strong> <strong>s<strong>of</strong>t</strong> <strong>classification</strong>.<br />

For example, entropy can be used (Zhang and Foody 1998). Entropy describes the variations<br />

in class membership probabilities associated with each pixel.<br />

The entropy is calculated by,<br />

∑<br />

H ( p)<br />

= − p(<br />

x)<br />

log 2 p(<br />

x)<br />

where p(x) is the class membership probabilities.<br />

But entropy is only suitable when ground <strong>data</strong> are hard. When both classified <strong>data</strong> and<br />

ground <strong>data</strong> are fuzzy, cross-entropy are more appropriate (Zhang and Foody 1998). Other<br />

possible indices <strong>of</strong> <strong>classification</strong> accuracy may be <strong>based</strong> on correlation analysis and distance<br />

measures between the representations <strong>of</strong> the land cover in the image <strong>classification</strong> and ground<br />

<strong>data</strong>.<br />

One other alternative is to degrade the <strong>data</strong> to just contain one or two classes in each pixel<br />

and use conventional methods, but the result is not an evaluation <strong>of</strong> the fuzzy <strong>classification</strong>.<br />

There are a number <strong>of</strong> points to note. Firstly, the predictive accuracy <strong>of</strong> partial values tends<br />

to vary between the land cover classes being considered (Bastin 1997). Secondly, the<br />

estimates <strong>of</strong> pixel composition give no reliable estimate <strong>of</strong> where in the pixel objects or<br />

features are likely to be.<br />

To deal with the second point, there are methods for sharpening the fuzzy <strong>classification</strong><br />

output (Foody 1998). This is made by fusion between a fine spatial resolution image and the<br />

fuzzy <strong>classification</strong> derived at a coarser resolution.<br />

There is a difficulty in evaluating <strong>s<strong>of</strong>t</strong> <strong>classification</strong>s, and there need to be done more<br />

research for better methods. There is no general measure that says when the <strong>classification</strong> is<br />

good or bad.<br />

25


4.5.4.2 Criteria and methods available<br />

One method that can be used along the whole continuum <strong>of</strong> <strong>classification</strong> fuzziness, from the<br />

<strong>case</strong> where every stage is hard to the <strong>case</strong> where every stage is fuzzy, is the RMS error (rootmean-square<br />

error) between the estimated and actual class composition <strong>of</strong> the pixels (Foody<br />

1999),<br />

where xi is a measurement, t is the true value and n is the number <strong>of</strong> <strong>data</strong> to be assessed.<br />

4.5.4.3 A suggested new approach<br />

One new measure can be to calculate how many pixels that are under- or overclassified a<br />

certain percent, in 10%- intervals, for each module.<br />

4.5.4.4 Implementation<br />

RMS =<br />

4.5.4.4.1 Validation <strong>data</strong> from reference <strong>data</strong><br />

In the evaluation RMS errors were calculated for all samplings and modules to see<br />

how well the <strong>classification</strong> performed in the classes. The training site pixels and<br />

the ‘unknown’ areas in the reference image, i.e. classes that had not<br />

been used, were excluded and then the whole images were used in the validation. To be<br />

able to compare the modules and samplings, a weighted total average RMS error was<br />

calculated. Each class’ RMS error was weighted with the reference area <strong>of</strong> the same class,<br />

Average_RMS_error =Σ RMS_error⋅ percentuel area in the reference image<br />

Also the average value for each class, in percent was calculated. This was been done in<br />

two ways, one compared to the reference image and one compared to the classified pixels.<br />

This to be able to compare BELclass with BAYclass and the reference image. BELclass<br />

underclassifies and may just classify 10% <strong>of</strong> the image.<br />

An analysis were made to see if the <strong>classification</strong> improved when more training pixels<br />

were used, in the sampling where the pixels were randomly taken, using BELclass, and<br />

how many classes that could be extracted in the same sampling.<br />

4.5.4.4.2 Statistical analysis<br />

I used Micro<strong>s<strong>of</strong>t</strong> Excel for the calculations.<br />

( x t)<br />

∑ i −<br />

i<br />

n<br />

26<br />

2


5. Results<br />

5.1 Introduction<br />

This chapter contains the results in terms <strong>of</strong> RMS errors, over- and underclassified pixels and<br />

average values.<br />

5.2 Fuzzy signatures<br />

Some <strong>of</strong> the eight chosen classes extracted have very similar spectral signatures, when using<br />

two near-infrared bands and one red band. That is why the <strong>s<strong>of</strong>t</strong> classifiers classify so many<br />

pixels, with different percentage <strong>of</strong> probability, to wrong classes. The mean values for each<br />

spectral band for each class, for the sampling in majority masks, is shown in figure 9. There it<br />

can be seen that coniferous, clear-cut forest, deciduous forest, mire and urban have mean<br />

values close to each other and therefore the training pixels can be mixed.<br />

Mean value<br />

120<br />

100<br />

80<br />

60<br />

40<br />

20<br />

0<br />

-20<br />

Fuzzy signatures for the six spectral bands in the majority sampling<br />

1 2 3 4 5 6<br />

Spectral bands (NIR,NIR,NIR,NIR,red,red)<br />

Figure 9. The mean values for the fuzzy signatures in the six spectral bands used ( 1-2 near<br />

infrared from September 1996, 3-4 near infrared from July 1995, 5 red from July 1995 and 6<br />

red from September 1996).<br />

5.3 RMS errors<br />

To be able to compare the different modules and samplings, each class’ RMS error were<br />

weighted with the percentage area <strong>of</strong> the reference image for respective class, to retrieve an<br />

overall RMS error, see table 2 below. This table shows that BAYclass performs best<br />

<strong>classification</strong> result before FUZclass. BELclass is not that far behind. They differ just a little<br />

between the modules and samplings but the result improves when the training pixels get purer<br />

27<br />

agri<br />

clear<br />

conif<br />

decid<br />

marsh<br />

mire<br />

urban<br />

water


and separable signatures are used. The modules classified differently well for the different<br />

classes, as seen in the more detailed tables, 3-5, showing the RMS errors for each class,<br />

sampling and module. In table 3 and 4, it can be seen that the RMS errors are reduced for the<br />

wetland classes when using a mask with areas that manually had been tried out to be<br />

separable. A thematic class can not <strong>of</strong>ten be assembled in one spectral class. This is why<br />

clustering or supervised <strong>classification</strong> need to be done to achieve separable classes within the<br />

masks.<br />

The number <strong>of</strong> classes you try to extract in BELclass makes no difference in the RMS<br />

errors. This is because BELclass uses unknown classes. It is only the spectral signature that<br />

counts. In that way one class is independent to another class, see table 4. More classes in<br />

BAYclass on the other hand improves the RMS errors for most classes.<br />

Table 2. Overall RMS errors for eight classes, weighted with the area <strong>of</strong> each class in the<br />

reference image, for the different modules and methods (×100 for percent).<br />

Overall RMS for the different samplings, named 1-4:<br />

1. training sites within 2. within majority 3. in majority mask and with 4. randomly taken<br />

module aggregated mask masks wetland signatures trainingsites<br />

BAYclass 0.299 0.259 0.244 0.362<br />

BELclass 0.404 0.400 0.398 0.436<br />

FUZclass 0.304 0.283 0.258 0.373<br />

Table 3. RMS errors for BAYclass.<br />

RMS for 8 classes classified with BAYclass RMS for 12<br />

classes trainingsites<br />

within<br />

aggregated<br />

mask<br />

within majority<br />

masks<br />

28<br />

in majority mask randomly taken<br />

classes,<br />

BAYclass<br />

randomly<br />

and with<br />

training sites taken training<br />

wetland signatures sites<br />

water 0.13 0.11 0.11 0.23 0.18<br />

urban 0.16 0.15 0.15 0.19 0.13<br />

agri 0.23 0.23 0.22 0.35 0.4<br />

marsh 0.15 0.16 0.07 0.13 0.11<br />

decid 0.21 0.21 0.22 0.24 0.24<br />

conif 0.42 0.33 0.3 0.48 0.5<br />

mire 0.21 0.2 0.12 0.2 0.19<br />

clear 0.22 0.22 0.24 0.21 0.2<br />

rock 0.16<br />

barr 0.13<br />

peat 0.07<br />

grass 0.08


Table 4. RMS errors for BELclass.<br />

RMS for 8 classes classified with BELclass RMS for 12<br />

29<br />

classes<br />

classes training sites within majority in majority mask randomly randomly<br />

within aggregatedmasks and with wetlandtaken taken<br />

masks signatures training sites training sites<br />

water 0.19 0.20 0.20 0.23 0.22<br />

urban 0.11 0.11 0.11 0.13 0.13<br />

agri 0.41 0.41 0.41 0.46 0.46<br />

marsh 0.08 0.08 0.03 0.08 0.08<br />

decid 0.24 0.24 0.24 0.27 0.27<br />

conif 0.56 0.55 0.55 0.59 0.59<br />

mire 0.12 0.12 0.06 0.13 0.13<br />

clear 0.22 0.22 0.22 0.22 0.22<br />

rock 0.03<br />

barr 0.06<br />

peat 0.04<br />

grass 0.06<br />

Table 5. RMS errors for FUZclass.<br />

RMS for 8 classes classified with FUZclass<br />

classes training sites within within majority in majority mask and with randomly taken<br />

aggregated masks masks wetland signatures training sites<br />

water 0.13 0.11 0.11 0.22<br />

urban 0.13 0.13 0.15 0.18<br />

agri 0.22 0.21 0.18 0.37<br />

marsh 0.16 0.15 0.04 0.16<br />

decid 0.22 0.21 0.21 0.24<br />

conif 0.44 0.40 0.36 0.50<br />

mire 0.16 0.18 0.09 0.15<br />

clear 0.21 0.22 0.24 0.21<br />

5.4 Over- and under classified pixels<br />

When looking at the number <strong>of</strong> correct classified pixels (± 5% ) in the summaries <strong>of</strong> table 6<br />

and 7, it can be seen that BELclass is superior with an average <strong>of</strong> 80%, compared to<br />

BAYclass’ 63%. But in these tables, the values are not weighted with the area, and the small<br />

classes that don’t contain so many pixels therefore brings up the average value, since they are<br />

so well classified with BELclass. The under- and overestimated values for each class in the<br />

sum-up tables 6 and 7, is the percent in pixels, not the total RMS error. The detailed tables for<br />

each 10% interval are shown as tables 8 and 9.


Table 6. Percent pixels over- and underestimated, for BELclass (×100 for percent).<br />

Over-, correct- or underestimated for BELclass, method 2<br />

water urban agri marsh decid conif mire clear<br />

underestimated: 0.082 0.027 0.285 0.011 0.255 0.585 0.044 0.161<br />

correct +-5%: 0.918 0.968 0.708 0.970 0.702 0.409 0.929 0.808<br />

overestimated: 0.001 0.005 0.006 0.019 0.042 0.007 0.027 0.030<br />

average:<br />

underest. 0.181<br />

correct 0.802<br />

overest. 0.017<br />

Table 7. Percent pixels over- and underestimated pixels for BAYclass (×100 for percent).<br />

Over-, correct- or underestimated for BAYclass, method 2<br />

water urban agri marsh decid conif mire clear<br />

underestimated: 0.026 0.014 0.140 0.009 0.187 0.486 0.029 0.109<br />

correct +- 5%: 0.907 0.848 0.820 0.823 0.522 0.354 0.423 0.341<br />

overestimated: 0.067 0.139 0.040 0.169 0.291 0.160 0.548 0.550<br />

average:<br />

underest. 0.125<br />

correct 0.630<br />

overest. 0.245<br />

BELclass systematically underclassifies, as seen in table 8 or sum-up table 6. Figure 10 and<br />

11 visualizes the clear-cut forest class from table 8 and 9 to easier see how many pixels that<br />

are over- and underclassifed for BAYclass and BELclass and how much.<br />

30


Table 8. Percent pixels over- and underclassified for BELclass. For example 5.1% <strong>of</strong><br />

agricultural areas are underestimated between 84-95%.<br />

RMS in 10 % intervals for BELclass, method 2<br />

water urban agri marsh decid Conif mire clear<br />

100-94% underestimated 0.012 0.004 0.070 0.002 0.004 0.099 0.004 0.010<br />

95-84% 0.0130.0030.051 0.001 0.008 0.096 0.003 0.009<br />

85-74% 0.009 0.002 0.036 0.001 0.013 0.077 0.002 0.011<br />

75-64% 0.008 0.0030.030 0.001 0.020 0.067 0.004 0.013<br />

65-54% 0.009 0.0030.025 0.001 0.027 0.064 0.004 0.017<br />

55-44% 0.007 0.002 0.020 0.001 0.027 0.047 0.004 0.014<br />

45-34% 0.006 0.002 0.013 0.001 0.024 0.030 0.003 0.012<br />

35-24% 0.004 0.002 0.012 0.001 0.031 0.032 0.004 0.017<br />

25-14% 0.005 0.002 0.012 0.001 0.040 0.031 0.006 0.022<br />

15-4% 0.009 0.002 0.016 0.002 0.060 0.0430.010 0.037<br />

5% underest. - 5% overest. 0.918 0.968 0.708 0.970 0.702 0.409 0.929 0.808<br />

6-15% overestimated 0 0.0030.004 0.009 0.025 0.004 0.021 0.016<br />

16-25% 0 0.001 0.001 0.004 0.010 0.002 0.005 0.007<br />

26-35% 0 0 0 0.003 0.004 0 0.001 0.003<br />

36-45% 0 0 0 0.001 0.002 0 0 0.001<br />

46-55% 0 0 0 0.001 0.001 0 0 0.001<br />

56-65% 0 0 0 0 0 0 0 0.001<br />

66-75% 0 0 0 0 0 0 0 0<br />

76-85% 0 0 0 0 0 0 0 0<br />

86-95% 0 0 0 0 0 0 0 0<br />

96-100% 0 0 0 0 0 0 0 0<br />

Procent pixlar<br />

Procent över- eller underklassade pixlar i hyggesklassen, klassad med BELclass<br />

0,900<br />

0,800<br />

0,700<br />

0,600<br />

0,500<br />

0,400<br />

0,300<br />

0,200<br />

0,100<br />

0,000<br />

100-94% underestimated<br />

95-84%<br />

85-74%<br />

75-64%<br />

65-54%<br />

55-44%<br />

45-34%<br />

35-24%<br />

25-14%<br />

15-4%<br />

5% underest. - 5% overest.<br />

6-15% overestimated<br />

16-25%<br />

26-35%<br />

36-45%<br />

46-55%<br />

56-65%<br />

66-75%<br />

76-85%<br />

86-95%<br />

96-100%<br />

10 % - intervall för över- och underklassat<br />

Figure 10. Percent under- and overclassified pixels for the clear-cut forest class with<br />

BELclass.<br />

31<br />

hyggen


Percent pixels<br />

Table 9. Percent pixels over- and underclassified for BAYclass. For example 1.4 % <strong>of</strong><br />

agricultural areas are underestimated between 84-95%.<br />

Over- or underclassified, in 10 % intervals for BAYclass, sampling 2<br />

water urban agri marsh decid conif mire clear<br />

100-94% underestimated 0.001 0 0.012 0 0 0.001 0 0<br />

95-84% 0.001 0.001 0.014 0 0.002 0.005 0.001 0.002<br />

85-74% 0.001 0.001 0.012 0 0.004 0.014 0.001 0.005<br />

75-64% 0.002 0.001 0.0130 0.009 0.035 0.0030.010<br />

65-54% 0.0030.002 0.014 0.001 0.015 0.060 0.003 0.012<br />

55-44% 0.0030.001 0.013 0.001 0.022 0.081 0.004 0.015<br />

45-34% 0.002 0.001 0.011 0.001 0.025 0.083 0.004 0.015<br />

35-24% 0.003 0.002 0.013 0.001 0.031 0.080 0.004 0.015<br />

25-14% 0.004 0.002 0.015 0.001 0.036 0.069 0.004 0.016<br />

15-4% 0.007 0.0030.023 0.0030.044 0.058 0.005 0.019<br />

5% underest. - 5% overest. 0.907 0.848 0.820 0.823 0.522 0.354 0.423 0.341<br />

6-15% overestimated 0.036 0.081 0.014 0.087 0.152 0.074 0.217 0.172<br />

16-25% 0.011 0.018 0.009 0.025 0.070 0.0430.1730.226<br />

26-35% 0.008 0.009 0.006 0.013 0.038 0.022 0.094 0.121<br />

36-45% 0.005 0.006 0.005 0.009 0.018 0.011 0.036 0.026<br />

46-55% 0.002 0.005 0.002 0.008 0.009 0.005 0.014 0.004<br />

56-65% 0.002 0.004 0.002 0.005 0.004 0.0030.006 0<br />

66-75% 0.001 0.0030.001 0.005 0.001 0 0.003 0<br />

76-85% 0.001 0.0030.001 0.006 0 0 0 0<br />

86-95% 0 0.005 0.001 0.007 0 0 0 0<br />

96-100% 0 0.005 0 0.004 0 0 0 0<br />

1.000<br />

0.900<br />

0.800<br />

0.700<br />

0.600<br />

0.500<br />

0.400<br />

0.300<br />

0.200<br />

0.100<br />

0.000<br />

Percent pixels under- or overclassified in the 'clear' class with BAYclass<br />

100-94% underestimated<br />

95-84%<br />

85-74%<br />

75-64%<br />

65-54%<br />

55-44%<br />

45-34%<br />

35-24%<br />

25-14%<br />

15-4%<br />

5% underest. - 5% overest.<br />

6-15% overestimated<br />

16-25%<br />

26-35%<br />

36-45%<br />

46-55%<br />

56-65%<br />

66-75%<br />

76-85%<br />

86-95%<br />

96-100%<br />

Percent under- or overclassified intervals<br />

Figure 11. Percent under- and overclassified pixels for the clear-cut forest class with<br />

BAYclass.<br />

32<br />

clear


5.5 Average values<br />

If you look at the average value, in percent <strong>of</strong> the reference image, for each class classified<br />

with BELclass, in table 10, you see that no sampling performs very well. This can be due to a<br />

low amount <strong>of</strong> training sites. It is a limitation in the <strong>s<strong>of</strong>t</strong>ware that only allow 10 000 random<br />

points. The small classes need <strong>of</strong> up to 40 000 points to get about 100 training pixels for each<br />

class in the aggregated mask. So probably the result improves if the training sites are collected<br />

manually and separable signatures are extracted. BAYclass have some higher occurence and<br />

comes therefore closer to the true average value for some classes, but overclassifies for many<br />

classes as well. It is not so strange that BAYclass have higher occurence, since this module<br />

think that the trained classes are the only one existing. If there are other classes in the image,<br />

it puts them to the most look-alike class. BELclass on the other hand treats the pixel as<br />

belonging to an unknown class if the pixel doesn’t support any <strong>of</strong> the given classes.<br />

As seen in average table 11, BELclass classifies more <strong>of</strong> the image when more classes are<br />

added, 13% compared to 0.90% for eight classes. But it is in wrong classes the increase is<br />

done.<br />

33


Table 10. Average values for 8 classes in percent <strong>of</strong> reference image and percent <strong>of</strong> classified<br />

image.<br />

Average values for 8 classes (in %) Average values for<br />

classified with BELclass for the different samplings BAYclass and FUZclass<br />

for one sampling<br />

class reference training sites within in majority mask randomly BAYclass FUZclass<br />

image within % <strong>of</strong> majority % <strong>of</strong> and with wetland % <strong>of</strong> taken % <strong>of</strong> in majority in majority<br />

aggregated classified mask, classified signatures, classified training classified mask, mask,<br />

mask, area % <strong>of</strong> ref. area % <strong>of</strong> ref. area sites area % <strong>of</strong> ref. % <strong>of</strong> ref.<br />

% <strong>of</strong> ref.<br />

% <strong>of</strong> ref.<br />

water 6.03 0.98 7.88 0.78 6.90 0.75 6.44 0.03 2.8 5.81 6.16<br />

urban 2.11 0.37 2.95 0.35 3.10 0.35 3.03 0.38 42.6 5.71 9.23<br />

agri 23.3 4.04 32.57 2.98 26.46 3.15 27.21 0.00 0.4 15.47 17.03<br />

marsh 0.82 0.82 6.65 0.46 4.12 0.01 0.08 0.36 39.9 6.35 7.40<br />

decid 13.68 2.37 19.11 2.23 19.77 2.47 21.33 0.00 0.3 12.74 12.35<br />

conif 42.71 2.84 22.93 3.07 27.23 3.18 27.47 0.00 0.0 25.18 18.30<br />

mire 2.81 0.46 3.75 0.46 4.13 0.06 0.51 0.06 6.9 13.54 14.25<br />

clear 8.53 0.52 4.16 0.93 8.29 1.61 13.94 0.06 7.1 15.20 15.28<br />

Sum %: 100.00 12.40 100.00 11.26 100.00 11.58 100.00 0.90 100.0 100.00 100.00<br />

35


Table 11. Average values for 12 classes in percent <strong>of</strong> reference image and percent <strong>of</strong><br />

classified image.<br />

Average values (in %) for 12 classes<br />

classified with BELclass and BAYclass for sampling 4<br />

class reference BELclass,<br />

BAYclass,<br />

image random training sites<br />

random training sites<br />

% <strong>of</strong> ref. % <strong>of</strong> classified % <strong>of</strong> classified area<br />

area<br />

water 5.67 0.16 1.16 12.07<br />

urban 1.99 0.01 0.04 5.98<br />

agri 21.93 0.00 0.00 5.68<br />

marsh 0.77 7.81 56.84 5.23<br />

decid 12.87 0.00 0.00 7.09<br />

conif 40.19 0.00 0.01 10.39<br />

mire 2.65 0.00 0.02 13.86<br />

clear 8.03 0.01 0.08 8.72<br />

rock 0.89 0.02 0.12 13.61<br />

barr 0.69 3.47 25.22 6.62<br />

peat 0.48 1.43 10.42 4.97<br />

grass 3.83 0.84 6.08 5.80<br />

Sum %: 100.00 13.74 100.00 100.00<br />

As can be seen in table 12, the number <strong>of</strong> training pixels does not matter when they are<br />

taken randomly. That is because the fuzzy partitioning matrix in the <strong>s<strong>of</strong>t</strong>ware doesn’t allow<br />

a not-square matrix and therefore an average value is calculated for each class.<br />

Table 12. Comparison <strong>of</strong> RMS errors with 3000<br />

training sample points and 10 000.<br />

RMS for 8 classes classified with BELclass<br />

classes randomly taken randomly taken<br />

trainingsites trainingsites<br />

RESURS 3000 RESURS 10000<br />

water 0.23 0.23<br />

urban 0.13 0.13<br />

agri 0.46 0.46<br />

marsh 0.08 0.08<br />

decid 0.27 0.27<br />

conif 0.59 0.59<br />

mire 0.13 0.13<br />

clear 0.22 0.22<br />

In figure 12-37 the reference images for all classes along with the result for each class for<br />

BAYclass and BELclass are shown for visual comparison. The agricultural class, the<br />

deciduous-, coniferous and clear-cut forest classes are labelled with shortenings in the figure<br />

texts.<br />

36


Figure 12. The agri.class<br />

from the reference image.<br />

Figure 15. The water class<br />

from the reference image.<br />

Figure 18. The urban class<br />

from the reference image.<br />

Figure 13. The agri.class<br />

classified with BAYclass.<br />

Figure 16. The water class<br />

classified with BAYclass.<br />

Figure 19. The urban class<br />

classified with BAYclass.<br />

37<br />

Figure 14. The agri.class<br />

classified with BELclass.<br />

Figure 17. The water class<br />

classified with BELclass.<br />

Figure 20. The urban class<br />

classified with BELclass.


Figure 21. The mire class<br />

from the reference image.<br />

Figure 24. The decid class<br />

from the reference image.<br />

Figure 27. The clear class<br />

from the reference image.<br />

Figure 22. The mire class<br />

classified with BAYclass.<br />

Figure 25. The decid class<br />

classified with BAYclass.<br />

Figure 28. The clear class<br />

classified with BAYclass.<br />

38<br />

Figure 23. The mire class<br />

classified with BELclass.<br />

Figure 26. The decid class<br />

classified with BELclass.<br />

Figure 29. The clear class<br />

classified with BELclass.


Figure 30. The conif class<br />

form the reference image.<br />

Figure 33. The marsh class<br />

from the reference image.<br />

Figure 31. The conif class<br />

classified with BAYclass.<br />

Figure 34. The marsh class<br />

classified with BAYclass.<br />

Figure 36. Uncertainty<br />

image for BAYclass.<br />

39<br />

Figure 32. The conif class<br />

classified with BELclass.<br />

Figure 35. The marsh class<br />

classified with BELclass.<br />

Figure 37. Uncertainty<br />

image for BELclass.


6. Discussion<br />

The reason why the water is surprisingly bad classified in all tryouts, can be due to error in<br />

the reference image. One lake in the reference image is practically covered with reed in the<br />

RESURS images. Training pixels in that lake can reduce the training site average for the<br />

class, and make at least BELclass to not classify water pixels if to far away from the mean<br />

value. BAYclass handles the problem a bit better since it only chooses among the classes<br />

given. Coniferous forest and agricultural areas have the highest RMS errors. This can be due<br />

to the fact that these classes are not spectrally homogenous. Coniferous pixels can also have<br />

been classified as water, since water and coniferous are the darkest pixels in the image. One<br />

class that is surprisingly well classified is the urban class. This can be due to that time series<br />

are used.<br />

There are many reasons why the classifiers perform unsatisfactory. The major reason is<br />

the input <strong>data</strong>. The RESURS images only had three, or actually only two spectral bands with<br />

good enough radiometric quality. These were the two near infrared and the red bands.<br />

Therefore, two images from different times were used, to at least have six bands. But in one<br />

<strong>of</strong> these images the spectral bands did not fit the other so a geometric correction had to be<br />

done. All images were then geometrically corrected to each other and to the reference image.<br />

There can still be small errors in the geometry, which affects the <strong>classification</strong> and the<br />

evaluation.<br />

The reference <strong>data</strong> is not really true either. It has been developed using topographic maps to<br />

get the boundaries <strong>of</strong> the classes and then for some classes just classified within the mask.<br />

Some other classes have been extracted using GIS-operations.<br />

The reference image is <strong>based</strong> on interpretation, GIS-operations and maximum-likelihood<br />

<strong>classification</strong>, why the true percentage distribution is not really known. What is really<br />

between 50-100% is set to be 100%. So the assumptions about what really is 100% may not<br />

be true. The classes, which are mixed in the reference image, as for example mixed forest, are<br />

not considered in the validation. The content <strong>of</strong> each forest kind in a pixel is unknown. But<br />

they have not participated in the training phase either.<br />

40


7. Conclusions and recommendations<br />

The advantages <strong>of</strong> <strong>s<strong>of</strong>t</strong> classifiers are that small classes will not vanish as with maximumlikelihood<br />

<strong>classification</strong> and they gives a measure, not in whole pixels, <strong>of</strong> the occurrence <strong>of</strong><br />

the classes.<br />

To receive an acceptable <strong>classification</strong> result, the training areas need to be spectrally<br />

separable. This can be done with clustering or expert knowledge. There is also needed to be<br />

collected enough number <strong>of</strong> training areas to get the spectral variation <strong>of</strong> each class.<br />

If only red and near infrared bands are available, a time series is required to get enough<br />

information. The images can be chosen close in time to get rid <strong>of</strong> clouds and more important<br />

from different vegetation periods for a better <strong>classification</strong>. The <strong>satellite</strong> images also needs to<br />

have good geometric and radiometric quality.<br />

Forest and agricultural areas are hardest to classify because <strong>of</strong> the heterogeneous spectral<br />

signatures. Urban on the other hand is surprisingly well classified. This can be due to the time<br />

series. BAYclass is the <strong>s<strong>of</strong>t</strong> classifier that gives the best result.<br />

There need to be done some improvements in the <strong>s<strong>of</strong>t</strong>ware to be able to use any <strong>of</strong> these<br />

<strong>s<strong>of</strong>t</strong> classifiers in production. The main drawback is that the fuzzy partition matrix has to be<br />

square. So even if you take training sites with separable signatures manually, you have to<br />

calculate an average value for all your training sites for a certain class. This reduces the value<br />

<strong>of</strong> many training sites.<br />

The second drawback, if still wanting the computer to take training sites, is that not enough<br />

training sites can be taken for small classes. Not enough randomly spread out pixels will hit<br />

the areas where a member <strong>of</strong> a small class is situated.<br />

The third drawback if wanting to use images from different dates, is that the maximum<br />

number <strong>of</strong> spectral bands allowed is seven when using IDRISI.<br />

When these improvements are done, these <strong>s<strong>of</strong>t</strong> classifiers can be a valuable tool when using<br />

medium- or coarse resolution <strong>satellite</strong> <strong>data</strong> to extract more information from the images.<br />

41


References<br />

Arora, M.K. and Foody, G.M. 1997, Log-linear modelling for the evaluation <strong>of</strong> the variables<br />

affecting the accuracy <strong>of</strong> probabilistic, fuzzy and neural network <strong>classification</strong>.<br />

International Journal <strong>of</strong> Remote Sensing 4:785-798<br />

Augusteijn, M.F. and Warrender, C.E. 1998, Wetland <strong>classification</strong> using optical and radar<br />

<strong>data</strong> and neural network <strong>classification</strong>. International Journal <strong>of</strong> Remote Sensing 8:1545-<br />

1560<br />

Bastin, L. 1997, Comparison <strong>of</strong> fuzzy c-means <strong>classification</strong>, linear mixture modelling and<br />

MLC probabilities as tools for unmixing coarse pixels. International Journal <strong>of</strong> Remote<br />

Sensing 17:3629-3648<br />

Bråkenhielm, S., Drakenberg, B., Hellgren, M., Holmgren, P., Karltun, E., Melkerud, P-A.<br />

and Olsson, M. 1998, Markinfo. Institutionen för skoglig marklära, SLU http://wwwmarkinfo.slu.se<br />

(visited October the 15 th 1999)<br />

Eastman, J.R. 1997, IDRISI for Windows Version 2.0: User´s Guide. Clark Labs, Clark<br />

University, Worcester, Main, USA<br />

euromap Satellitendaten- Vertriebsgesellschaft mbH, 1997, IRS-1C Handbook.<br />

http://www.euromap.de/ (visited October the 10 th 1999)<br />

Foody, G.M. 1996, Approaches for the production and evaluation <strong>of</strong> fuzzy land cover<br />

<strong>classification</strong> from remotely-sensed <strong>data</strong>. International Journal <strong>of</strong> Remote Sensing 7:1317-<br />

1340<br />

Foody, G.M. 1998, Sharpening fuzzy <strong>classification</strong> output to refine the representation <strong>of</strong><br />

sub-pixel land cover distribution. International Journal <strong>of</strong> Remote Sensing 13:2593-2599<br />

Foody, G.M. 1999, The continuum <strong>of</strong> Classification Fuzziness in Thematic Mapping.<br />

Photogrammetric Engineering & Remote Sensing 4:443-451<br />

Foody, G.M. and Boyd, D.S. 1999 Fuzzy mapping <strong>of</strong> tropical land cover along an<br />

environmental gradient from remotely sensed <strong>data</strong> with an artificial neural network.<br />

Journal <strong>of</strong> Geographical Systems 1:23-35<br />

Fuller, D. and Fulk, M. 1998, An Assessment <strong>of</strong> Fire Distribution and Impacts during 1997<br />

in Kalimantan, Indonesia Using Satellite Remote Sensing and Geographic Information<br />

Systems. The George Washington University and Clark University<br />

Gualtieri, J.A. and Withers, J. 1988, Goddard researchers simulate neural networks using<br />

parallel processing. NASA Information Systems Newsletter 15:12-15<br />

Howald, K.J. 1989, Neural network image <strong>classification</strong>. Proceedings ASPRS-ACSM Fall<br />

Convention from Compass to Computer, Cleveland 207-215<br />

Hyyppä, J. 1999, WP 3200 State <strong>of</strong> the art in EO mapping, monitoring and <strong>data</strong> accessibility:<br />

Report on Earth Observation methods and <strong>data</strong> accessibility <strong>of</strong> mediumresolution<br />

<strong>satellite</strong> <strong>data</strong>. BALANS – Planning and management in the Baltic Sea Region<br />

with land information from EO, Environment and Climate Programme, Area 3.3: Centre for<br />

Earth Observation.<br />

Kanellopoulos, I., Varfis, A., Wilkinson, G.G. and Mégier, J. 1991, Classification <strong>of</strong><br />

Remotely-Sensed Satellite Images Using Multi-Layer Perceptron Networks in Proceedings<br />

<strong>of</strong> the 1991 International Conference on Artificial Neural Networks (ICANN-91), Espoo,<br />

Finland (Kohonen, T., Mäkisara, K., Simula, O. and Kangas, J.,eds.) Elsevier Science<br />

Publications (North-Holland) 2:1067-1070<br />

42


Kanellopoulos, I. 1997, Use <strong>of</strong> Neural Networks for Improving Satellite Image Processing<br />

Techniques for Land Cover/Land Use Classification, SUPCOM'95 LOT 16. Provision <strong>of</strong><br />

Statistical Services – Support to the Commission EUROSTAT Remote Sensing and<br />

Statistics Programme. Institute for Systems, Informatics and Safety (Neural Networks<br />

Laboratory) & Space Applications Institute (EMAP Unit - Advanced Methods Sector),<br />

European Commission, Joint Research Centre I-21020 Ispra (VA), Italy<br />

http://ams.egeo.sai.jrc.it/eurostat/Lot16-SUPCOM95/ (visited September the 15 th 1999)<br />

Langaas, S. 1996, Transboundary European GIS <strong>data</strong>bases: A review <strong>of</strong> the Baltic Sea<br />

region experiences, GISDATA Specialist Meeting on Geographic Information: the<br />

European Dimension. BALANS – Planning and management in the Baltic Sea Region with<br />

land information from EO, Environment and Climate Programme, Area 3.3: Centre for<br />

Earth Observation.<br />

Lillesand, T.M. and Kiefer, R.W. 1994, Remote Sensing and Image Interpretation. John<br />

Wiley & Sons, Inc. 587-618<br />

Malmberg, U. and Furberg, O. 1997 Klassnings- och kompositmetoder för medelupplösande<br />

satellit<strong>data</strong>, Pilotstudie med RESURS-<strong>data</strong> över centrala Östersjöområdet. Swedish Space<br />

Corporation, Rymdbolaget X-PUBL-11<br />

Mannan, B. and Roy, J. and Ray, A.K. 1998, Fuzzy ARTMAP supervised <strong>classification</strong> <strong>of</strong><br />

multi-spectral remotely-sensed images. International Journal <strong>of</strong> Remote Sensing 4:767-774<br />

Manyara, C.G. and Lein, J.K. 1994, Exploring the Suitability <strong>of</strong> Fuzzy Set Theory in Image<br />

Classification: A Comparative Study Applied to the Mau Forest Area Kenya.<br />

ASPRS/ACSM, http://www.odyssey.maine.edu/gisweb/spatdb/acsm/ac94044.htm<br />

(visited September the 16 th 1999)<br />

Maselli, F. And Rodolfi, A. and Conese, C. 1996, Fuzzy <strong>classification</strong> <strong>of</strong> spatially degraded<br />

Thematic Mapper <strong>data</strong> for the estimation <strong>of</strong> sub-pixel components. International Journal<br />

<strong>of</strong> Remote Sensing 3:537-551<br />

Olsson, B. 1999, BALANS Land Cover Information for the Baltic Sea Drainage Basin.<br />

BALANS – Planning and management in the Baltic Sea Region with land information from<br />

EO, http://www.ssc.se/rst/rss/BALANS/ (visited October th 15 th 1999)<br />

Palubinskas,G. 1994, Post-processing <strong>of</strong> fuzzy <strong>classification</strong> <strong>of</strong> forests for Landsat TM<br />

imagery. EGIS Foundation<br />

Pouncey, R. and Flynn, T. and Curry, K. 1997, ERDAS IMAGINE Ver. 8.3.1 On-Line<br />

Manuals. ERDAS, Inc.<br />

Ryan, T.W., Sementelli, P.J., Yuen, P. and Hunt, B.R. 1991, Extraction <strong>of</strong> shoreline features<br />

by neural nets and image processing. Photogrammetric Engineering and Remote Sensing<br />

7:947-955<br />

SSC Rymdbolaget, 1998, Produktspecifikation av landtäcke<strong>data</strong> version 1.0. SSC<br />

SSC Satellitbild, 1999, RESURS-O1 Technical Specifications.,<br />

http://www.ssc.se/sb/RESURS/technic.html (visited October the 11 th 1999)<br />

Zhang, J. and Foody, G.M. 1998, A fuzzy <strong>classification</strong> <strong>of</strong> sub-urban land cover from<br />

remotely sensed imagery. International Journal <strong>of</strong> Remote Sensing 14:2721-2738<br />

43


Appendix 1: Action steps for <strong>s<strong>of</strong>t</strong> <strong>classification</strong><br />

This is an example for water.<br />

Actions Input Module Output<br />

Make a water mask reference_gripen RECLASS rcwater<br />

in 12.5 m pixelsize<br />

Aggregate to a pixel rcwater CONTRACT aggwater<br />

size <strong>of</strong> 150 m<br />

Scale the pixel values aggwater SCALAR scwater<br />

to be in range 1-100,<br />

but still real values<br />

Convert the pixel values scwater CONVERT cvwater<br />

to byte binary<br />

For method 2 and 3:<br />

Mask out 75-100 % <strong>of</strong> cvwater RECLASS majwater<br />

a class, 90-100% for cvdecid… majdecid…<br />

water<br />

Convert the majority majwater CONVERT cv2water<br />

masks to integer binary majdecid… cv2decid…<br />

Take out sample points majwater SAMPLE vecwater<br />

randomly to be majdecid… vecdecid…<br />

training sites from the<br />

mask.<br />

Initialize images with aggwater INITIAL trswater<br />

the same parameters as trsdecid…<br />

an aggregated image<br />

The initialized image vecwater POINTRAS trswater<br />

updated with rasterized vecdecid… trsdecid…<br />

sample points<br />

Reclass the sample trswater RECLASS rc2water<br />

points to 1 and 0 trsdecid… rc2decid…<br />

Overlay to extract just majwater, rc2water OVERLAY ovwater<br />

the sample points which majdecid, rc2decid… ovdecid…<br />

are within the mask<br />

44


Numbering the sample ovwater GROUP grwater<br />

points ovdecid… grdecid…<br />

Reclass the sample grwater RECLASS grwater1<br />

points for one class to grdecid… grdecid1…<br />

only one classnumber<br />

Make an 8-layer image, grwater1 OVERLAY gr1_8<br />

(if you have 8 classes), grdecid1…<br />

with the classes (sample<br />

point groups).<br />

Extract an average value grwater1, aggwater, EXTRACT valwater<br />

for each class in each grdecid1, aggdecid… valurban…<br />

class mask.<br />

Create a fuzzy partitioning matrix. In Database Workshop<br />

Number 8 rows for the use <strong>of</strong> 8 classes. -Number the rows. One more<br />

Name 8 columns with class names, row comes up with the down<br />

e.g water, decidious. arrow. To create one more<br />

column:<br />

-Modify<br />

-Add field<br />

Real (4 byte) and class name<br />

Put in the values in the valwater In Database Workshop<br />

matrix valurban… -File<br />

-import/export Idridi values file<br />

free format ASCII (VAL file)<br />

valwater… and corresponding<br />

class<br />

Link the matrix to the gr1_8, water In Database Workshop fzwater<br />

training sites (sample gr1_8, decid… -Assign field values to fzdecid….<br />

points groups) image gr1_8 fzwater..<br />

and corresponding class<br />

Create the fuzzy gr1_8, sig.names, FUZSIG water.sig<br />

signatures <strong>satellite</strong>bands decid.sig….<br />

Classify with different signature names MAXLIKE mlclass<br />

hard/<strong>s<strong>of</strong>t</strong> classifiers BAYCLASS baywater…<br />

BELCLASS belwater…<br />

FUZCLASS fuzwater…<br />

45


Calculate the over- baywater, aggwater OVERLAY errwater<br />

and underclassified baydecid, aggdecid… errdecid…<br />

pixel values<br />

Add 1 to the extracted errwater SCALAR sc2water<br />

error-values and errdecid… sc2decid…<br />

multiply with 100 to<br />

have positive values<br />

between 1 and 200<br />

Convert the pixel values sc2water CONVERT cv3water<br />

to byte binary sc2decid… cv3decid…<br />

Export the error images and <strong>classification</strong> images to Imagine to receive pixel information:<br />

Export the images to cv3water, baywater -File expwater.img<br />

ERDAS Imagine cv3decid… - Export ex2water.img<br />

-Export<br />

-S<strong>of</strong>tware-Specific Formats<br />

-ERDIDRIS<br />

-IDRISI to ERDAS 7.4<br />

name <strong>of</strong> file to export and name in<br />

Imagine with extension .img<br />

In Imagine:<br />

-Open the image<br />

-info i<br />

-compute statistics<br />

In the viewer:<br />

-Raster<br />

-Attributes<br />

-Edit<br />

Mark the column<br />

-Copy<br />

Copy the pixel information into excel to calculate RMS and average values for each class<br />

and each module.<br />

In Excel:<br />

-Paste the number <strong>of</strong> pixels in the excel-file “mall_RMS.xls” and “mall_medeltal.xls”<br />

46


Appendix 2: Macr<strong>of</strong>ile- scheme<br />

8 classes, aggregated mask<br />

NEWMASK<br />

8 classes, majority mask<br />

MACMAJ_3<br />

MACMAJ_5<br />

MACR_8_2<br />

TRSITES3<br />

MACR8_4<br />

MCAGG8_5<br />

8 classes, miresig, aggregated mask<br />

NEWMASK2<br />

MACR8_6 , create matrix<br />

FUZSIG (FUZSIG4 if Landsat TM)<br />

MAC8_BAY MAC8_BEL MAC8_FUZ MAC8_ML<br />

Export to Imagine and Excel. eval.. for RMS for random<br />

cv10.. for average for random<br />

cv4.. for RMS other methods<br />

cv3.. for average other methods<br />

47<br />

8 classes, random<br />

MAC_RD_2 , create matrix<br />

FUZSIG2 (FUZSIG3 IF Landsat TM)


TRITA-GEOFOTO serien utges av:<br />

Kungl Tekniska Högskolan<br />

Institutionen för Geodesi och Fotogrammetri<br />

100 44 STOCKHOLM<br />

Drottning Kristinas väg 30<br />

Telefon: 08-7907342<br />

Telefax: 08-7907343<br />

URL: http://www.geomatics.kth.se/<br />

TRITA-GEOFOTO-serien 1999 - The TRITA-GEOFOTO series, 1999<br />

1999:1 Friborg, Klas. Differential GPS in A Tested Mobil Environment – An Accuracy Study.<br />

March 1999. (Examensarbete i geodesi nr 3061. Handledare: Horemuz)<br />

1999:2 Torlegård, Kennert. Analytisk fotogrammetri och dess felteori. (Kompendium)<br />

1999:3 Horemuz, Milan and Lars E Sjöberg. Quick GPS Ambiguity Resolution For Short<br />

And Long Baselines. March 1999.<br />

1999:4 Anna-Maria Dorfh och Maria Nilsson. Digitalt ort<strong>of</strong>oto för ajourhållning av<br />

baskarta. April 1999. (Examensarbete i fotogrammetri nr 112. Handledare: Anders<br />

Boberg)<br />

1999:5 Rose-Marie Kvål och Annica Larsson. Algal Bloom Monitoring and Detection in the<br />

Baltic Sea, the Usefulness <strong>of</strong> Marine Sensors Compared to a Meteorological Sensor. April<br />

1999. (Examensarbete i fotogrammetri. Handledare: Boberg, Lundén och Rud)<br />

1999:6 Krzyszt<strong>of</strong> Gajdamowicz. Automated Processing <strong>of</strong> Georeferenced Colour Stereo Images<br />

for Road Inventory. April 1999. (Doctoral dissertation)<br />

1999:7 Daniel Öhman. Automatic Text Localisation on Traffic Signs from Colour Stereoscopic<br />

Image Sequences. April 1999. (Examensarbete i fotogrammetri nr 113. Handledare: Boberg,<br />

Gajdamowicz)<br />

1999:8 Stefan Wulcan. Samlingskarta - En förstudie avseende samordning av ledningar inom<br />

Solna. May 1999. (Examensarbete. Handledare: Hauska)<br />

1999:9 Magnus Johansson och Christer Bylund. Landhöjning i södra Sverige<br />

beräknad från tredje precisionsavvägningen och RH70 höjder. Juni 1999. (Examensarbete i<br />

geodesi nr 3062. Handledare: Fan)<br />

1999:10 Jan Danielsen. Determination <strong>of</strong> Land Uplift in areas not covered by repeated levellings.<br />

With application to South-Norway. Juni 1999. (Licentiate thesis)<br />

1999:11 Mattias Pettersson. Utvärdering av vektoriseringsprogram. Juni 1999.<br />

(Examensarbete i geoinformatik. Handledare: Hauska)<br />

1999:12 Anders Vidén. Semiautomatiserad fotogrammetrisk mätning av signalerade punkter, linjeoch<br />

ellipsobjekt i MIDOC, Multi Image Documentation.". Juni 1999. (Examensarbete i<br />

fotogrammetri nr 114. Handledare: Axelsson)<br />

1999:13 Albert Morales Pérez and Pere Carmona Ortega. Movements monitoring <strong>of</strong> a high<br />

building. Juni 1999. (Examensarbete i geodesi nr 3063. Handledare: Horemuz)


1999:14 Lina Nyberg och Andreas Prevodnik. A Pilot Study <strong>of</strong> Soil Erosion in the Close<br />

Environment <strong>of</strong> Idrija, Western Slovenia. An Application <strong>of</strong> Geographical Information<br />

Systems. December 1999. (Examensarbete i geoinformatik. Handledare: Henkel)<br />

1999:15 Sanna Sparr Olivier. Évaluation de la possibilité d'utiliser des images<br />

<strong>satellite</strong>s à haute résolution et le GPS différentiel pour la cartographie en<br />

haute montagne - Avec réalisation d'un prototype. Augusti 1999.<br />

(Examensarbete i fotogrammetri nr 115. Handledare: Boberg)<br />

1999:16 Rannala, Marek. Surveying <strong>of</strong> road geometry with Real Time Kinematics GPS. Oktober<br />

1999. (Examensarbete i geodesi nr 3064. Handledare: Horemuz)<br />

1999:17 Klang, Dan. Reconstruction <strong>of</strong> Geometric Road Data Using Remotely Sensed Imagery.<br />

Oktober 1999. Photogrammetric Reports No 67. (Doctoral Dissertation)<br />

TRITA-GEOFOTO-serien 2000 - The TRITA-GEOFOTO series, 2000<br />

2000:1 Carballo, Gabriel. Statistically-<strong>based</strong> multiresolution network flow phase unwrapping in<br />

SAR interferometry. January 2000. Photogrammetric Reports No 68. (Doctoral Dissertation)<br />

2000:2 Jonas Andersson och Jens Hedlund. Geografiska analyser över Internet. Januari 2000.<br />

(Examensarbete i geoinformatik. Handledare: Roslund)<br />

2000:3 Thomas Rehders. Noggrannhetsstudier vid RTK-mätning och kvalitetsundersökning av GPSmottagare.<br />

Februari 2000. (Examensarbete i geodesi nr 3065. Handledare: Horemuz)<br />

2000:4 Åsa Hesson. Publicering av historiska kartor på Internet. Februari 2000. (Examensarbete i<br />

geoinformatik. Handledare: Hauska)<br />

2000:5 Malin Skagerström och Karin Wiklander. Extension <strong>of</strong> a control network in dar es salaam<br />

using GPS-technology. Februari 2000. (Examensarbete i geodesi nr 3066. Handledare:<br />

Stoimenov)<br />

2000:6 Maria Sjödin och Anna Strid. Automatisk generalisering med objektorienterad kart<strong>data</strong>bas.<br />

Februari 2000. (Examensarbete i geoinformatik. Handledare: Hauska)<br />

2000:7 Per Koldestam. IT-projektet Ortnamnslinjen. Mars 2000. (Examensarbete i geoinformatik.<br />

Handledare: Hauska).<br />

2000:8 Anna Källander och Elin Söderström. Deformationsanalys i horisontal- och vertikalled på<br />

Slussen-konstruktionen. Mars 2000. (Examensarbete i geodesi nr 3067. Handledare: Horemuz<br />

och Lewen)<br />

2000:9 Anna Haglund. <strong>Towards</strong> <strong>s<strong>of</strong>t</strong> <strong>classification</strong> <strong>of</strong> <strong>satellite</strong> <strong>data</strong> - A <strong>case</strong> <strong>study</strong> <strong>based</strong> upon Resurs<br />

MSU-SK <strong>satellite</strong> <strong>data</strong> and land cover <strong>classification</strong> within the Baltic Sea Region. Mars 2000.<br />

(Examensarbete i geoinformatik. Handledare: Roslund och Langaas).

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!