Sie sind auf Seite 1von 19

Classification Accuracy Assessment

Dr. Sao Hone Pha


Director and Head
Remote Sensing and GIS Research Center
Yangon Technological University

Date: 15th, October 2018


Certificate Training Course on Basic Concept of Remote Sensing & GIS (1.11.2018) to (30.11.2018)
2

Accuracy Assessment
Accuracy
The degree (often expressed as a percentage) of correspondence between observation and reality.

We usually judge accuracy against existing maps, large scale aerial photos, or field checks
(information used as reference, considered “truth”).

 Comparison to two sources of information


 Remote Sensing derived classification map
 Reference Test information
 The relationship between the two sets of information's is expressed as a matrix known as
 ERROR MATRIX / CONFUSION MATRIX / CONTINGENCY TABLE
3

Classification Error Matrix


Error matrices compare, on a category-by-category basis, the relationship between known
reference data (ground truth) and the corresponding results of an automated classification.

Square matrix with the number of rows and columns equal to the number of categories
whose classification accuracy is being assessed.

Error matrix stems from classifying the sampled training set pixels and listing the known cover
types for training (columns) versus the pixel actually classified into each land cover category
by the classifier (rows).
4

Several Characteristics about Classification Performance

Major diagonal of the matrix  The training set pixels that are classified into the proper
and cover categories

All nondiagonal elements  Errors of Omission of the matrix (or)Errors of Commission


5

Errors
Omission Error (related to “producer accuracy”)

 Error of exclusion
 Pixels are not assigned to its appropriate class, or
Pixels are omitted from the actual class they belong
 Correspond to non diagonal column element

Commission Error (related to “user accuracy”)

 error of inclusion
 pixel is assigned to a class to which it does not belong,
(or) pixels were improperly included in that category
 represented by non diagonal row element
6

ERROR of Omission Vs. Commission


(From a Class1 Perspective)

Commission error: pixels in Class2 Omission error: pixels in Class1


erroneously assigned to Class 1 erroneously assigned to Class 2

Class 1 Class 2
Number of
pixels

0 Digital Number 255


7

ERROR of Omission Vs. Commission


(From a Class2 Perspective)

Omission error: pixels in Commission error: pixels in


Class2 erroneously assigned to Class1 erroneously assigned to
Class 1 Class 2

Class 1 Class 2
Number of
pixels

0 Digital Number 255


8

Overall Accuracy =
The total number of correctly classified pixels
The total number of reference pixels

Accuracy of individual category =


The number of correctly classified pixel in each category
The total number of pixels in the corresponding row or column
9

Producer’s accuracy = (1-Omission error)


Omission error of a class = 100 – (% Accuracy of a class)

Producer Accuracy  indicates how well training set pixels of the given cover types are classified
The number of correctly classified pixels in each category (major diagonal)
The number of training set pixels used for that category (column total)

User Accuracy  indicates the probability that a pixel classified into a given category actually represented
that category on the ground
The number of correctly classified pixels in each category (major diagonal)
The total number of pixels that were classified in that category (row total)

User’s Accuracy = (1-Commission Error)


Commission error of a class = 100 – (% Accuracy of a class)
10

So, from this assessment we have three measures of accuracy which address subtly different
issues:

Overall accuracy :

takes no account of source of error (errors of omission or commission)

User accuracy :
measures the proportion of each TM class which is correct.

Producer accuracy

measures the proportion of the land base which is correctly classified.


11

Kappa Coefficient ( 𝒌𝒌 or “KHAT”)

K statistics is a measure of the difference between the actual agreement between the
reference data and an automated classifier and the chance agreement between the reference
data and a random classifier.


Conceptually 𝑘𝑘 can be defined as

Observed accuracy - chance agreement


k =
ˆ
1 - chance agreement
12

Kappa Coefficient ( 𝒌𝒌 or “KHAT”)

Observed accuracy - chance agreement


k =
ˆ
1 - chance agreement

The statistic serves as an indicator of the extent to which the percentage correct values of an
error matrix are due to the “true” agreement versus “chance” agreement

A true agreement (observed accuracy) approaches 1 and chance agreement approaches 0


then Kappa approaches 1

A kappa of 0 suggest that the classification is very poor and the given classifier is no better
than a random assignment of pixels
13

∑x
i =1
ii

Observed Accuracy =
N
r

1 ∑ (x i+ ⋅ x+ i )
Change Agreement = i =1
N N

Where
r = number of rows in the error matrix
xii = number of observations in row i and column i (on the major diagonal)
xi+ = total of observations in row i (shown as marginal total to right of the matrix)
x+i = total of observations in column i (shown as marginal total at bottom of the matrix)
N = total number of observations included in matrix
14

The KHAT statistic is computed as

Observed accuracy - chance agreement


k =
ˆ
1 - chance agreement
r r

∑x ii
1 ∑ (x i+ ⋅ x+ i )
i =1
- i =1

N N N
k̂ = r

1 ∑ (x i+ ⋅ x+ i )
1 - i =1
N N
r r
N ∑ xii − ∑ ( xi + ⋅ x+ i )
kˆ = i =1
r
i =1

N − ∑ ( xi + ⋅ x+ i )
2

i =1
15

• Note : Overall Accuracy is biased on the correctly classified pixels (the major
diagonal) and excludes the omission and commission errors

• Kappa in addition to the correctly classified pixels, includes the off diagonal elements
as a product of Row total and column total
16
CLASSIFIED IMAGE REFERENCE DATA

From airphotos, field data, etc.

1 Generate a random sample of points

2 For each point, compare the map class with the “true” class

3 Construct and analyze an error matrix...


17

Error Matrix Resulting from Classifying Random ly Sam pled Test


Reference Data

Classification Data Water Sand Forest Urban Corn Hay Row total

Water 226 0 0 12 0 1 239

Sand 0 216 0 92 1 0 309

Forest 3 0 360 228 3 5 599

Urban 2 108 2 397 8 4 521

Corn 1 4 48 132 190 78 453

Hay 1 0 19 84 36 219 359

Column total 233 328 429 945 238 307 2480

Producer’s accuracy User’s accuracy

W = 226/233 = 97% U = 397/945 = 42% W =226/239= 94% U = 397/521 = 76%

S = 216/328 = 66% C= 190/238= 80% S = 216/309 = 70% C = 190/453 = 42%

F = 360/429 = 84% H = 219/307 =71% F = 360/599 =60% H = 219/359 = 61%

Overall accuracy = (226+ 216+ 360 + 397+ 190+219)/2480= 65%


18

Error Matrix Resulting from Classifying Random ly Sam pled Test

r r
N ∑ xii − ∑ ( xi + ⋅ x+ i )
kˆ = i =1
r
i =1

N 2 − ∑ ( xi + ⋅ x+ i )
i =1

∑x
i =1
ii = 226 + 216 + 360 + 397 + 190 + 219 = 1608

∑ (x
i =1
i+ ⋅ x+ i ) = (239 × 233) + (309 × 328) + (599 × 429)

+ (521× 945) + (453 × 238) + (359 × 307) =


1,124, 382

(2480 × 1608) − 1,124, 382


kˆ = 0.57
(2480) 2 − 1,124, 382
19

How accurate should the classified map be?

General rule of thumb is 85% accuracy

What’s a good Kappa?

General range
K < 0.4: poor
0.4 < K < 0.75: good
K > 0.75: excellent

Das könnte Ihnen auch gefallen