• Nu S-Au Găsit Rezultate

Evaluation of Image Segmentation

N/A
N/A
Protected

Academic year: 2022

Share "Evaluation of Image Segmentation"

Copied!
31
0
0

Text complet

(1)

Evaluation of Image Segmentation

1 Introduction

This technical report describes several performance measures that are used in literature in order to evaluate the image segmentation.

2 Scientific problem

The segmentation problem:

• two-labels: one class of interest and one class of background;

• multiple labels: more classes of interest and one class of background (e.g. in tumour segmentation, core and edema represents two class of interest; taking into account the background class, a 3-label segmentation problem is obtained).

The segmentation of an image,I, can be defined as the partitioning ofI inLsubregions,R1,R2,. . .,RL, such that:

• ∪nL

l=1Rl=I

∀ex∈Rl, S(ex) =l,∀l∈ {1,2, . . . , L}

Rlis a connected set,∀l∈ {1,2, . . . , L}

Rl1 ∩Rl2 =ϕ,∀l1, l2, l1 ̸=l2

Q(Rl) =T rue, l∈ {1,2, . . . , L}

Q(Rl1∪Rl2) =F alsefor each pair of adjacent regionsRl1,Rl2.

whereQis a logical predicate defined on the points of the considered region (Qis used for characterizing the objects of the image).

In the case of two regions (L = 2: background and foreground)Rb andRf (from a segmentation S), an element is defined to be on the boundary ofRb andRf if at least one of its connected neighbours belongs to a different category (Rf orRb). The boundary elements in an image form a setBS.

3 Evaluation of segmentation

In order to evaluate the quality of segmentation algorithms, various methodologies are possible. The criteria that make differences between these practices regard

• the reference segmentation to that the computed segmentation is compared or

• the working principles of the evaluation methods [Bauer et al., 2009].

(2)

3.1 Reference segmentation

In the first case, the aim of evaluation is to compare (by using similarity measures or distance-based mea- sures) the computed segmentations to the segmentations created by human experts [Niessen et al., 1998], [Gerig et al., 2001]. Even if the segmentations produced by humans (medical experts) could not represent a true gold standard, this framework is the most objective one.

A second case regards a ranking system in which the computed segmentations receive ranks, given by human experts. Such approach is subjective and requires numerous experts and a large time in order to analyse all the images and to rate them [Shimizu and Nawano, 2004].

A third case is represented by a comparison of segmentation methods by taking into account their common agreement (in [Warfield et al., 2004] the STAPLE algorithm is suggested). Even if this evaluation provides good results, some risks appear (e.g., the algorithms obtain good scores by simply producing the same errors as other methods, which also leads to a high common agreement).

3.2 Working principles

Another criteria for classifying the evaluation methods regards the working principles of them [Cardenes et al., 2007], [Cárdenes et al., 2009]. Taking into account if a human analyses (in a visual way) or not the produced segmen- tation, the evaluation methods can be classified as:

1. subjective methods — that require the human evaluator; the quality of these methods depend upon the standard used by different experts in order to evaluate the segmentations or upon the order in which evaluators observe the segmentation results;

2. objective methods — that do not require the human evaluator. By taking into account the impact of segmentations, these objective methods can be classified as:

(a) System-level evaluation methods — the segmentation method is analysed by involving it into an overall system; the performance of the system will highlight the segmentation quality;

(b) Direct evaluation methods — the segmentation method is analysed independently. By considering the object of examination, the direct evaluation methods are classified as:

i. analytical evaluation methods — based only on intrinsic features of the segmentation algo- rithms (e.g. processing strategy (parallel, sequential, iterative, or mixed), processing complex- ity, resource efficiency, and segmentation resolution); these methods are aimed to analyse an algorithm in terms of principles, requirements, complexity etc. without reference to a con- crete implementation of the algorithm or test data. For example, one can define the time complexity of an algorithm, or its response to a theoretical data model such as a step edge [Zhang et al., 2008].

ii. empirical evaluation methods — based on the study of the results

A. discrepancy (or supervised) methods – compare the results with a reference image or gold standard or ground truth; measures as the number of misclassified voxels, position of mis- classified voxels, number of objects in the image or feature values of segmented objects could be used in order to evaluate the quality of segmentation.

• based on the number of misclassified pixels versus the reference image, with penalties weighted proportional to the distance to the closest correctly classified pixel for that region [10-14].

• based on the differences in the feature values measured from the segmented images and the reference image [15-20]. These methods have been extended to accommodate the problem when the number of objects differs between the segmented and reference images [21-25]

(3)

• evaluation of edge-based image segmentation methods [26-28,68,30-35]

• Everingham et al. [36] proposed a method to comprehensively evaluate segmentation algorithms using the Pareto front. Instead of using a single discrepancy metric and evalu- ating effectiveness in a discrepancy space, it performs evaluation in a multi-dimensional fitness/cost space with multiple discrepancy metrics

B. goodness (stand-alone or unsupervised) methods – based on the study of the results them- selves; evaluate a segmented image based on how well it matches a broad set of charac- teristics of segmented images as desired by humans. These methods evaluate algorithms by computing a “goodness” metric on the segmented image without a priori knowledge of the desired segmentation result.

4 Supervised evaluation

It is very important to establish how we define similar regions or segmentations. The obtained segmentations and their boundaries could be compact, discontinuous, smooth, etc. In order to define a similarity measure several aspects have to be considered: colour, texture, motion. When a distance metric is involved in the evaluation process, the minimum, the mean or the maximum of such a measure could be considered.

4.1 Prerequisites

Firstly, we give the definitions of four basic cardinalities of the so-called confusion matrix, namely the true positives (TP), the false positives (FP), the true negatives (TN), and the false negatives (FN) for crisp segmen- tations.

• 2D images: LetI(ex) : R2 →Rbe a 2D (medical) image (composed by pixels), andS(ex) :R2 Ω, Ω ={0,1,2, . . . , L1}, be anL-ry decision segmentation of the imageI(ex).

• 3D images: LetI(x) :e R3 Rbe a 3D (medical) image (composed by voxels), andS(x) :e R3 Ω, Ω ={0,1,2, . . . , L1}, be anL-ry decision segmentation of the imageI(ex).

Suppose that we have two segmentations:

• the reference segmentation (gold standard or ground truth)1-Sr

• the computed segmentation (by a specific algorithm) -Sc. Each image point (pixel/voxel) can be classified as followed:

• true positive: (Sr(x)e is 1(Sc(x)e is 1),

• false positive:(Sr(ex)is 0∧(Sc(ex)is 1),

• true negative: (Sr(x)e is 0(Sc(x)e is 0),

• false negative:(Sr(x)e is 1(Sc(x)e is 0).

Each of these segmentations are composed by k segments/regions/classes (e.g. if k = 2, then the two segments are represented by the class of interest and the background; if k = 3, then two classes of interest and the background will represent the possible segments).

In the case of k = 2 segments, the confusion matrix can be represented as that from Table1.

1in order to call the reference segmentation “ground truth” we have to be certain that it is. Manual reference segmentations drawn by experts normally approximate ground truth, in which case it can be used as gold standard, but not as the ground truth itself.

(4)

real segments positive (interest)

segment

negative (background)

segment computed

segments

positive (interest)

segment TP FN

negative (background)

segment FN TN

Table 1: confusionMatrix 4.2 Measures based on spatial overlap

These metrics compute the overlap between regions. They quantify the similarity of two segmentations and they are useful when the volume changes are of importance for the image analysis.

4.2.1 Dice coefficient

• Definition: This similarity coefficient [Dice, 1945] is computed as the ratio between the number of pixels/voxels belonging to the intersection (of two possible segmentations) and the average of their sizes.

CoeffDice(Sc, Sr) = 2|Sr∩Sc|

|Sr|+|Sc| = 2T P

2T P +F P +F N (1)

• u.m.:Percent

• Range:Its values range between 0 (no overlap) and 1 (perfect agreement).

• Utility:It is able of quantifying the reproducibility (repeatability).

• Other names: It is also called the overlap index. A mathematical equivalent of Dice coefficient is the Fβ measure (defined by eq.2, withβ= 1))

Fβ = (β2+ 1)PrecisionRecall

β2Precision+Recall (2)

Another possible reference to Dice coefficient is the term Symmetric Volume Overlap (SVO) [Campadelli et al., 2009, Campadelli et al., 2008,Campadelli et al., 2010,Casiraghi et al., 2009], defined as 1 - the symmetric vol-

ume difference (SVD).

SV D= 1CoeffDice= 12∗ |Sc∩Sr|

|Sc|+|Sr| (3)

Coef fDice is a special case of kappa statistics if the number of true negatives is significantly larger than the other three statistical parameters.

• Code: Visceral

(5)

4.2.2 Jaccard coefficient

• Definition: The Jaccard index (JAC) [Jaccard, 1912] is computed as the ratio between the intersection (of two possible segmentations) and their union.

CoeffJaccard(Sc, Sr) = |Sr∩Sc|

|Sr∪Sc| = T P

T P +F P +F N (4)

• u.m.:Percent

• Range:Its values range between 0 (no overlap) and 1 (perfect agreement).

• Utility: In [Taha and Hanbury, 2015] is noted that CoeffJaccardand CoeffDiceare equal at the extrema 0, 1 and CoeffDice> CoeffJaccardbetween those limits. Furthermore the two metrics are related according to:

CoeffJaccard= CoeffDice

2CoeffDice.

Due to this relation, both coefficients evaluate the same aspects of segmentation, without providing sup- plementary information if both of them are used as validation metrics.

• Other names:Volumetric overlap error is the corresponding error measure [Ruskó et al., 2009] (V OE = 1CoeffJaccard).

• Code: Visceral

Dice and Jaccard coefficients are sensitive to misplacement of the segmentation label, but they are rela- tively insensitive to volumetric under- and overestimations. Shape infidelity is only captured if the deviation is volumetrically impactful: a thin panhandle will not result in a large deviation from one. The Dice coefficient is currently more popular than the Jaccard coefficient, since Jaccard is numerically more sensitive to mismatch when there is reasonably strong overlap. Dice values “look nicer” because they are higher for the same pair of segmentations. A drawback of both is that they are unsuitable for comparing segmentation accuracy on objects that differ in size [Rohlfing et al., 2004].

4.2.3 True positive rate&co

• Definition: Recall is computed as the ratio between the number of positive pixels/voxels in the reference image and the number of pixels/voxels identified as positive in the segmented image.

Recall= T P

T P +F N (5)

In conjunction with Precision (defined asT PT P+F P), Recall is used in order to compute F-measure.

Specificity is computed as the ration between the number of negative pixels/voxels in the reference image and the number of pixels/voxels identified as negative in the segmented image.

Specificity= T N

T N+F P (6)

These two measures depend on the size of segments. There are two other measures that are related to these metrics, namely Fallout and the false negative rate (FNR). They are defined by:

Fallout= F P

F P +T N = 1Specificity (7)

(6)

FNR= F N

F N+T P = 1Recall (8)

• u.m.: Percent

• Range:[0,1]

• Utility: Since the last two measures are equivalent to Specificity and Recall, only one pair ((Recall, Specificity) or (Fallout, False Negative rate)) should be used in order to evaluate and to analyze the performance of segmentation.

• Other names: Recall is also called Sensitivity or True Positive Rate. Specificity is also called True Negative Rate. Fallout is also called the false positive rate (FPR).

• Code: Visceral

4.2.4 Global consistency error

Another frequently used measure for evaluating the segmentation performance is the Global Consistency Error (GCE) [Martin et al., 2001]. An error-based measure is actually an “opponent” to a similarity measure (two segmentations are identical if an error-based measure is 0).

• Definition: This measure is computed as an average over the error of pixels/voxels belonging to two segmentations. It is able of comparing partitions of the same image and it is tolerant of one partition refining the other (e.g. by splitting regions or merging them together). First of all, if we consider an imageI (that containsnpixels/voxels: n= |I|) and a segmentation region of it S. We can note the set of all points (pixels/voxels) that are neighbours topi and that belong to the same segmentation regionS byR(S, pi). For two segmentationsScandSr, the (not symmetric) error (called local refinement error in [Martin et al., 2001]) at point (pixel/voxel)pi, LRE(Sc, Sr, pi)is defined as

LRE(Sc, Sr, pi) = |R(Sc, pi)−R(Sr, pi)|

|R(Sc, pi)| (9)

The GCE between segmentations can be defined as a mean over the error of all points (pixels/voxels):

GCE(S1, S2) = 1

|I|min



|I|

i=1

LRE(S1, S2, pi),

|I|

i=1

LRE(S2, S1, pi)



 (10) By using the cardinalities introduced in Section4.1, GCE can be expressed as follows:

GCE(Sc, Sr) = 1

|I|min

{F N(F N+ 2T P)

T P +F N +F P(F P + 2T N)

T N +F P ,F P(F P + 2T P)

T P +F P +F N(F N + 2T N) T N +F N

}

(11)

• u.m.:Percent

• Range:Its values range between 0 (perfect agreement) and 1 (most different).

• Utility: This measure is able of quantifying the consistency between image segmentations of differing granularities. It has the advantage of being tolerant of (label) refinement. Using this measure has sense when the two segmentations being compared have similar number of segments [Unnikrishnan et al., 2007].

• Code: Visceral

(7)

4.2.5 False positive Dice and False Negative Dice

Another possibility to analyse two segmentations is to use the False Positive Dice (FPD) and False Negative Dice (FND) measures [Babalola et al., 2009].

• Definition: False Positive Dice (FPD) is defined by Eq. 12and False Negative Dice (FND) is defined by Eq.13.

FPD(Sc, Sr) = 2Sc∩Sr

|Sc|+|Sr| (12)

FND(Sc, Sr) = 2Sc∩Sr

|Sc|+|Sr| (13)

whereScandSrare the complements of the segmentation results and gold standards, respectively.

• u.m.: Percent

• Range: [0,1]

• Utility:

• Other names: The FPD helps to evaluate the over-segmentation, while the FND gives a measure of under-segmentation.

• Code: none

4.3 Measures based on distance

These metrics compute the distance between two segmented regions (surfaces/volumes) by taking into account the point (pixel/voxel) location. They quantify the dissimilarity of two segmentations and they are useful when the contours (the shapes of the boundary of the structure) are of importance for the image analysis.

A distance value of 0 corresponds to a perfect match between the computed segmentation and the ground truth, while greater values indicate higher errors.

4.3.1 Mean symmetric surface distance (MSSD)

The mean surface distance (MSSD) [Babalola et al., 2009] takes into account the error between the surfaces (boundaries) of two segmentations (ScandSr) by using distances between their boundary pixels/voxels.

• Definition: The border points (pixels/voxels) of segmentation and reference are determined. These are defined as those points (pixels/voxels) in the object that have at least one neighbour (from all the nearest neighbours) that does not belong to the object. For each point (pixel/voxel) in these sets, the closest point (pixel/voxel) in the other set is determined (using Euclidean distance and real world distances, so taking into account the generally different resolutions in the different scan directions). All these distances are stored, for border pixels/voxels from both reference and segmentation. The average of all these distances gives the mean symmetric absolute surface distance.

The distance surface is defined for a point (pixel/voxel). Ifpi is theithsurface point (pixel/voxel) onSc (pi∈BSc), the distance frompito the closest point (pixel/voxel) onSris:

d(pi, Sr) = min

pjSr

d(pi, pj) (14)

(8)

whered(pi, pj)is the Euclidean distance of the points (pixels/voxels) incorporating the real spatial res- olution of the image. The distance has associate a sign: the negative one means that a particular point (pixel/voxel) inScis within the surface defined bySr, while the positive values otherwise.

The mean symmetric surface distance (MSSD) is the average of all the distances from points on the boundary of computed segmentation to the boundary of ground truth and from points on the boundary of ground truth to the boundary of the computed segmentation, respectively:

MSSD(Sc, Sr) = 1

|BSc|+|BSr

 ∑

piBSc

d(pi, Sr) + ∑

pjBSr

d(pj, Sc)

 (15)

• u.m.: Millimeters

• Range: This measure is 0 for a perfect segmentation and greater the value of the distance, the worse the segmentation is.

• Utility: This distance provides a measure of the average mutual distance between edges of the two surfaces.

• Code: none

4.3.2 Hausdorff distance and Average Hausdorff distance

• Definition: In order to compute this distance, some preliminary concepts must be explained.

The distance between a pointpbelonging to a segmentationScand another segmentationSris defined as:

d(p, Sr) = min

piSr

d(p, pi) (16)

where∥.∥denotes some norm, e.g. Euclidean distance.

The directed Hausdorff distance between two segmentations (ScandSr) is given by the maximal distance from a point in the first segmentation to a nearest point in the other one

d(Sc, Sr) = max

piSc

d(pi, Sr) (17)

This distance is in general not symmetrical [Aspert et al., 2002]; d(Sc, Sr) can be called the forward distance andd(Sr, Sc)can be called the backward distance.

In order to compute the symmetrical Hausdorff distance between two segmentations, for each point in Sc the minimum distance to all points in Sr is computed and vice versa (for each pixel/voxel on the boundary ofScthere is guaranteed to be a voxel ofSrin a distance of at most HAD, and vice versa).

Then, the maximum over the set of minimum distances will represent SHD. The symmetrical Hausdorff distance [Hausdorff, 1914] is defined as follows:

SHD(Sc, Sr) = max{d(Sc, Sr), d(Sr, Sc)} (18) The symmetrical distance allows of better evaluation of the errors between segmentations (the computa- tion of a “onesided” error can lead to significantly underestimated distance values ).

(9)

The Average Hausdorff Distance (AHD) represents the mean of the directed Hausdorff distance for all the points of the first set.

da(Sc, Sr) = 1

|Sc|

piSc

d(pi, Sr) (19)

AHD(Sc, Sr) = max{da(Sc, Sr), da(Sr, Sc)} (20)

• u.m.:Millimeters

• Range:This measure is 0 for a perfect segmentation and greater the value of the distance, the worse the segmentation is.

• Utility: The Hausdorff distance is in fact the maximum symmetric surface distance. The HD is gen- erally sensitive to outliers. Because noise and outliers are common in medical segmentations, it is not recommended to use the HD directly [Zhang and Lu, 2004]. However, the quantile method proposed by Huttenlocher et al. [Huttenlocher et al., 1993] is one way to handle outliers. According to the Hausdorff quantile method, the HD is defined to be theqth quantile of distances instead of the maximum, so that possible outliers are excluded, where q is selected depending on the application and the nature of the measured point sets. The AHD is known to be stable and less sensitive to outliers than the HD.

It is sensitive to outliers and returns the true maximum error. This is required for applications as surgical planning, where the worst case error is more important than average errors.

• Other names:Also known as maximum symmetric surface distance.

• Code: Visceral

4.3.3 Mahalanobis distance

• Definition: The Mahalanobis Distance (MHD) [Mahalanobis, 1936] regards the correlation among all the points (pixels/voxels) from the set that the considered two points (pixels/voxels) are belonging.

MHD(p1, p2) =

(p1−p2)TCov1(p1−p2) (21) whereCovdenotes the covariance matrix associated to all the points (pixels/voxels) from the set.

If we want to compute the Mahalanobis distance between two sets, then the average over all elements of them must be considered.

MHD(Sc, Sr) =

Sc−µSr)TCov1Sc−µSr) (22) whereCovrepresents the covariance matrix associated to two segmentations.

• u.m.: Millimeters

• Range:This measure is 0 for a perfect segmentation and greater the value of the distance, the worse the segmentation is.

• Utility: Note that if Cov is an identity matrix, the Mahalanobis distance reduces to the Euclidean distance. This measure is able of automatically assigning weights based on statistical variation in the data.

It gives each dimension “equal” weight and also accounts for correlation between different dimensions.

• Code: Visceral

(10)

4.3.4 JCd

• Definition:In [Cárdenes et al., 2009] a new similarity measure is presented. It is based on the distance [Pichon et al., 2004] from the misclassified points (pixels/voxels):

d(r, Sc, Sr) =





0 ifr∈Sc∩Sr

minpiScr−pi ifr∈Sr−Sc minpjSrr−pj ifr∈Sc−Sr

(23)

JCd(Sc, Sr) = |Sc∩Sr|

|Sc∩Sr|+∑

id2(pi, Sc, Sr) +∑

jd2(pj, Sr, Sc) (24) wherepiare the misclassified points (pixels/voxels) ofScthat should be classified asSr,pjare the points (pixels/voxels) ofSrthat should be classified asSc.

• u.m.: Percent

• Range: [0,1]

• Utility: The new measure is aimed to penalize more those points (pixels/voxels) that are more distant from their class in the gold standard. This penalisation is performed by some weights associated to misclassified points (pixels/voxels). In [Cárdenes et al., 2009] for a point (pixel/voxel) pi, the square of Euclidian distance to the nearest point (pixel/voxel) of the same class to pi is used to weight each misclassified point (pixel/voxel).

• Code: none

4.3.5 Root mean square symmetric surface distance

• Definition: The root mean square symmetric surface distance (RMSD) [Bauer et al., 2009] is defined as:

RMSD(Sc, Sr) =

1

|BSc|+|BSr√ ∑

p1BSc

d2(p1, BSr) + ∑

p2BSr

d2(p2, BSc) (25)

whered(p, S)represents the distance defined by Eq.14.

• u.m.:Millimeters

• Range:This measures is similar to the mean symmetric surface distance, but stores the squared distances between the two sets of border points (pixels/voxels).

• Utility: The RMSD distance is highly correlated with the average distance, but has the advantage that large deviations from the true contour are punished stronger [Bauer et al., 2009].

• Code: none

4.4 Measures based on connectivity

• Definition:In [Cárdenes et al., 2009] a new measure is presented that takes into account the connectivity of the two given segmentations. Firstly, two sets of points (pixels/voxels) are connected if all the points (pixels/voxels) from a set has one or more neighbours (a given neighbourhood must be defined) in the

(11)

other set. Secondly, the number of connected components is computed for each set (ScandSr) and for each classc(class/classes of interest or background):NScc andNScr, respectively. Then, the connectivity coefficient for a classcis determined as:

CCc(Sc, Sr) = min{ NSc

c, NSc

r

} NSc

c+NSc

r

(26)

• u.m.:Percent

• Range:The range ofCCcis[0,1].

• Utility:???

• Code: none

4.5 Measures based on boundaries

• Definition:In [Cárdenes et al., 2009] a similarity measure, called boundary Jaccard Coefficient (BJC)) is introduced that is based on the boundaries of the segmentations. If we note byBSc the boundary of a segmentationS(ScorSr) for a classc(class/classes of interest or the background), BCJ is computed as:

BJCc(Sc, Sr) = BScc∩BScr

BScc∪BScr (27)

• u.m.:

• Range:

• Utility:

• Code: none

4.6 Measures based on volumes

These measures are useful when the segmentation purpose is to identify for changes in size. They are sensitive to mis-estimations of the segmented volume more than anything else. In the case of a single measurement, the volume error (two times the volume difference over the volume sum) can be used, while in the case of multiple measurements (when an average result is of interest) the absolute volume error can be used.

4.6.1 Volumetric similarity

• Definition:The similarity of two segmentations can be computed by taking into account their volumes.

In order to determine such a measure, a distance between two volumes must be defined (the similarity being 1 - this distance). A possible definition [Taha and Hanbury, 2015] for the volumetric distance is the ratio of the absolute volume difference and the sum of the compared volumes.

VS(Sc, Sr) = 1−||Scc| − |Src||

|Scc|+|Src| = 1 |F N−F P|

2T P +F P +F N (28)

• u.m.:Percent

• Range: [0,1]

(12)

• Utility:This similarity is based on the absolute volume, the overlap between segments is absolutely not considered (the volumetric similarity can have its maximum value even when the overlap is zero).

• Code: Visceral

4.6.2 Relative volume difference

• Definition: Another distance between the segmented volumes is their difference in size as a fraction of the size of the reference:

RVD(Sc, Sr) =±|Sc| − |Sr|

|Sr| (29)

• u.m.:Percent

• Range:[1,1]???

• Utility: This distance is not symmetric (it is not a metric; its values are signed numbers). If this distance is 0, this does not imply that the two segmentations are identical or they actually overlap with each other. Taking into account this remark, it is recommended [Bauer et al., 2009] to use this measure in combination with other measures. When more quality measure are used in order to evaluate two segmentation, the absolute value of this measure must be used (e.g. in a ranking/scoring system).

• Code: none

4.7 Measures based pair counting

In order to describe some pair-counting based metrics, the four basic classical cardinalities must be defined for the pair-counting framework.

We suppose that a set of pointsXis composed by two partitions andP is the set of (n

2 )

tuples formed by all pairs of objects (inX×X). Depending on the place in the partitions of each object of a tuple(p1, p2)∈P, four groups of tuples are possible, the size of each category beinga,b,c andd, respectively.

• ifp1andp2 are place in the same subset in both partitionsScandSr;

• ifp1andp2 are place in the same subset inSr, but in different subsets inSc;

• ifp1andp2 are place in the same subset inSc, but in different subsets inSr;

• ifp1andp2 are placed in different subsets in both partitionsScandSr.

The first and the last groups indicate an agreement (a+d), while the second and the third group indicate the disagreement (b+c) between the two partitions. More details about how these pair-counting cardinalities can be computed based on the classical cardinalities can be found in [Taha and Hanbury, 2015].

4.7.1 Rand Index (RI)

• Definition: A similarity measure able to evaluate both clusterings and classifications (because it is not based on labels) is the Rand Index (RI), proposed in [Rand, 1971].

RI(Sc, Sr) = a+b

a+b+c+d (30)

Another possibility to compute the Rand Index is to consider all the points (pixels/voxels) of the image (pi, i= 1, . . .∥I∥), the labels associated to these points by the two segmentations:lciandlir, respectively,

(13)

and to compute the ratio of the number of pairs of points having the same label relationship inScandSr: R(Sc, Sr) = 1

(I 2

) ∑

i,j,i̸=j

[If(

lci =lcj∧lir =ljr) +If(

lic̸=ljc∧lir̸=lrj)]

(31)

whereIf is the identity function and the denominator is the number of possible unique pairs among∥I∥ data points [Unnikrishnan et al., 2007]. In [Unnikrishnan et al., 2007] it is noted that the number of unique labels inScandSris not restricted to be equal.

• u.m.:Percent

• Range:[0,1]

• Utility: The main idea of Rand Index is to count the pairs of pixels that have compatible label relation- ships in the two segmentations to be compared [Unnikrishnan et al., 2007].

• Code: Visceral

4.7.2 Adjusted Rand Index (ARI)

• Definition:The Adjusted Rand Index (ARI), proposed by Hubert and Arabie [Hubert and Arabie, 1985]

, is a modification of the Rand Index that considers a correction for chance . It is given by:

ARI(Sc, Sr) = 2(ad−bc)

b2+c2+ 2ad+ (a+d)(b+c) (32)

• u.m.:Percent

• Range:[0,1]

• Utility: The adjusted index has the property of having expected value equal to 0 and maximum value of 1. Since the unadjusted Rand index was in the range [0, 1], the adjusted index can take on a wider range of values, increasing the sensitivity of the measure [Unnikrishnan et al., 2005,Unnikrishnan and Hebert, 2005].

• Code: Visceral

4.7.3 Structural Similarity index (SSIM)

For grey level images a well known image similarity measure is the Structural Similarity index (SSIM) [Wang et al., 2004]

which takes luminance, contrast and structure of two images A and B into account:

SSIM(A, B) = (2µAµB+C1)(2σAB+C2)

2A+µ2B+C1)(σA2 +σB2 +C2) (33) where the constants are set toC1 = (0.01×255)2andC2 = (0.03×255)2. The SSIM index is then applied locally using an11×11circular-symmetric Gaussian weighting function, and the mean over the image is used as the final similarity measure.

5 Comparison of statistical validation metrics

Popovic has introduced three criteria in order to compare two segmentation validation measures: bias, con- sistency and discriminancy. In order to compare two segmentation results R1 andR2 for a validation metric f(Sc, Sr), the following relation is introduced: if f is rating R1 over R2, then we say that: f(R1, Sr)

(14)

f(R2, Sr). In [Popovic et al., 2007] Λ is defined as the domain ofSc and Sr, andN as a total number of possible instances ofScinΛ.

Cf. [Popovic et al., 2007], we have:

• Bias: The segmentation validation metricsf(Sc, Sr)is biased if a pair{R1, Sr}and{R2, Sr}, such that f(R1, Sr) =f(R2, Sr)andR1 ≻R2exists.

• Consistency: Two segmentation validation metricsf(R, Sr)andg(R, Sr)are consistent if a pair{R1, Sr} and{R2, Sr}, such thatf(R1, Sr)≻f(R2, Sr)andg(R1, Sr)≺g(R2, Sr)does not exist.

• Discriminancy: Segmentation validation metricf(Sc, Sr)is more discriminating thang(Sc, Sr)if a pair {S1, Sr} and{S2, Sr}, such thatf(S1, Sr) ̸= f(S2, Sr)and g(S1, Sr) = g(S2, Sr) exists and a pair {S1, Sr}and{S2, Sr}such thatf(S1, Sr) =f(S2, Sr)andg(S1, Sr)̸=g(S2, Sr)does not exist.

Because validation metrics rarely have total consistency, probabilistic measures have been defined in [Popovic et al., 2007]:

• Degree of bias: If a segmentation validation metricsf(Sc, Sr)hasNbpairs{Si, Sr}and{Sj, Sr}, such thatf(Si, Sr) =f(Sj, Sr)andSi ̸=Sj, degree of bias is defined as:

B = Nb

N (34)

• Degree of consistency: If for two segmentation validation metrics f(Sc, Sr) andg(Sc, Sr)there exist Ncpairs{Si, Sr}and{Sj, Sr}, such thatf(Si, Sr) f(Sj, Sr)andg(Si, Sr) g(Sj, Sr), degree of consistency is defined as

C= N −Nc

N (35)

• Degree of discriminancy: If for two validation metrics f(Sc, Sr) andg(Sc, Sr) there existNf g pairs {Si, Sr} and {Sj, Sr}, such that f(Si, Sr) ̸= f(Sj, Sr) and g(Si, Sr) = g(Sj, Sr), and Ngf pairs {Si, Sr}and{Sj, Sr}, such thatg(Si, Sr) ̸=g(Sj, Sr)andf(Si, Sr) =f(Sj, Sr), degree of discrimi- nancy is defined as

D= Nf g Ngf

(36) In [Popovic et al., 2007] it is also noted that the value of degree of consistency is in the range[0,1], the degree of discriminancy is in the range[0,], the degrees of consistency are symmetrical while the degrees of discriminancy are reciprocal (Df g =Dgf1).

6 Unsupervised evaluation

Notation:

I - initial image

Sc- computed segmented image

FI(p)- feature of pixelpin imageI (gray-level, color value (RBG), others)

|I|- size ofI (measured in number of pixels)

The performance of algorithms in solving the image segmentation task have been evaluated taking into account four important criteria (originally proposed in [Haralick and Shapiro, 1985]:

(15)

• the regions must be uniform and homogeneous (respect to some image characteristic) – this criterion has been evaluated by computing various measures of intra-region uniformity.

• the adjacent regions must be different (respecting the same characteristic considering by the previous criterion) – this criterion has been evaluated by computing various measures of inter-region uniformity.

• the interior of each region must be simple and without holes – this criterion has been evaluated by computing various semantic measures.

• the boundaries of each region must be simple, not ragged (in fact, they must be spatially accurate) – this criterion has been evaluated by computing various semantic measures, also.

Because the evaluation of the last two criteria requires some semantic information about the images to be segmented, we will discuss only the first two criteria in what follows.

6.1 Intra-region uniformity

6.1.1 Measures based on colour error

Intra-region visual error Eintra One of the most intuitive criterion being able to quantify the quality of a segmentation result is the intra-region uniformity. Levine and Nazif defined a criterion that calculates the uniformity of a region characteristic based on the variance of this characteristic [Levine and Nazif, 1985].

Eintrarepresents the proportion of misclassified pixels in an image.

Eintra = 1 1

|I|

L l=1

sRl

[

FI(s)

tRl

FI(t) ]2

[ maxsRl

FI(s)min

sRl

FI(s)

]2 (37)

Characteristics:

• Range: not specified

• Optimisation direction: maximisation

• Other names: intra-region uniformity (max); intra-region disparity (min)

• Advantages: not specified

• Disadvantages: not specified

Internal and External Contrasts The contrast can be defined at many levels: between two points/pixels/voxels, between two regions or at the level of the entire image.

The contrast between two pointsc(p1, p2)of an imageI is defined as:

c(p1, p2) =|FI(p1)−FI(p2)| (38) At the level of a region, two contrasts are possible: internal contrast and external contrast. These contrasts take into account the internal and external contrast of the regions measured in the neighborhood of each pixel.

If we noteNr(p)the neighborhood of the point/pixel/voxelp(a neighbourhood of a pointpis a setNr(p) consists of all points q such thatdistance(p, q) r and the number r is called the radius of Nr(p)),F(p) the point/pixel/voxel feature (e.g. intensity) andFM AX the maximum feature (e. g. maximum intensity), the

(16)

contrast insideInternalContrastand outsideExternalContrastof the regionRl(withl∈1,2, . . . , L) are respectively:

InternalContrast(Rl) = 1

|Rl|

pRl

max

nNr(p)Rl

c(p, n) (39)

ExternalContrast(Rl) = 1

|BorderRl|

p∈Rl

nNmaxr(p),n /Rlc(p, n) (40) where|BorderRl|represent the length of the border of regionRl.

Internal contrast measures the uniformity of each region. Ii is defined as the averageMaxContrastin that region, whereMaxContrastis the largest luminance difference between a pixel and its neighboring pixels in the same region.

The contrast of a regionRlbelongs to[0,1]range and is defined as:

C(Rl) =





1 InternalContrast(Rl)

ExternalContrast(Rl), if0< InternalContrast(Rl)< ExternalContrast(Rl) ExternalContrast(Rl), ifInternalContrast(Rl) = 0

0, otherwise

(41)

The global contrast is defined as:

C(I) = 1

|I|

L l=1

|Rl|C(Rl) (42)

Characteristics ofC(I):

• Range:[0,1]

• Optimisation direction: maximisation

• Other names:

• Advantages: not specified

• Disadvantages: not correctly taking into account strongly textured regions.

Discrepancy Discrepancy (D) D measures the gray-level difference between the original image and the output image after segmentation. It was proposed to evaluate thresholding-based segmentation techniques that separate the foreground object from the background [Weszka and Rosenfeld, 1978].

D=∑

pI

(FI(p)−FSc(p))2 (43)

Characteristics:

• Range:[0,2552]

• Optimisation direction: maximisation

• Other names:

• Advantages: not specified

• Disadvantages: not specified

(17)

6.1.2 Measures based on squared colour error

Within-class varianceσw2 Intra-region uniformity measure, the within-class variance,σ2w, is the sum of the squared color error of the foreground object and the background, weighted by their respective sizes.

σw2 =

L l=1

|Rl|

|I| errF2(Rl) (44)

whereerr2F(Rl)represents the squared error colour of regionRland is defined as:

errF2(Rl) = ∑

p∈Rl

(F(p)−F(Rl))2 (45)

F(Rl) =

pRl

F(p)

|Rl| ,∀l= 1,2, . . . L (46) Characteristics:

• Range: not specified

• Optimisation direction: minimisation

• Other names:

• Advantages: not specified

• Disadvantages: not specified

Square error of colourD(I) D(I)is a measure of intra-region uniformity. In [Rosenberger and Chehdi, 2000]

D(I)is computed as the average squared color error of each region weighted by its size.

D(I) = 1 L

L l=1

|Rl|errF2(Rl)

|Sc| (47)

Characteristics:

• Range: not specified

• Optimisation direction: maximisation

• Other names: intra-region uniformity

• Advantages: not specified

• Disadvantages: not specified

Normalised uniformityN U The normalized uniformity measureN U is a region uniformity measure and is the normalized sum of the squared color error of the foreground object and the background.

N U = 1−err2F(Robj) +errF2(Rbackground)

Z (48)

N U was used in the context of an evaluation approach based on a set of segmentation measures that consti- tute a performance vector (P V). TheP V vector stores the factors characterizing the segmentation, including region uniformity, region contrast, line contrast, line connectivity, and texture [Levine and Nazif, 1985]. N U

(18)

improved uponP V by enhancing the region uniformity measure inP V to use a normalized region uniformity measure [Sahoo et al., 1988].

Characteristics:

• Range: not specified

• Optimisation direction: not specified

• Other names:

• Advantages: not specified

• Disadvantages: Defined for two regions only.

Gray-level uniformityU Levine and Nazif [Levine and Nazif, 1985] have included some weighting coeffi- cients in the formula of the uniformity for images with more then two regions. Gray-level uniformity measure, U, describes intra-region uniformity. For a gray-scale image,Uis a measure of the weighted sum of the squared gray-level error of each region.

U = 1

L l=1

err2F(Rll

Z (49)

where:

ωlis a weighting factor (it could be defined asωl= |R|Il||) and

Z is a normalisation factor. Z = (Fmax−Fmin)2, where theFmax andFmindenote the maximum and minimum feature of points/pixels/voxels in the test image, respectively.

Characteristics:

• Range: not specified

• Optimisation direction: not specified

• Other names:

• Advantages: not specified

• Disadvantages: not specified

Average squared colour error measures F, F and Q Methods F, F andQ are based on the average squared color error of each region. They involve different penalties in order to deal to the over-segmentation and under-segmentation problems.

Liu and Yang [Liu and Yang, 1994] have proposed an evaluation function, inspired by the qualitative cri- teria for good image segmentation established by Haralick and Shapiro [Haralick and Shapiro, 1985] 2 that requires no user-defined parameters. In [Borsotti et al., 1998] the authors have identified some limitations in this evaluation function, and propose two new functions (F andQ).

2A good segmentation must satisfy two properties:

• A region has to contain only one primitive (a texture or a constant gray level) to guarantee that there is no under-segmentation.

Thus, a region is characterized by the stability of the statistics inside it.

• Two adjacent regions have to contain two different primitives to guarantee that there is no over-segmentation. This corresponds to a disparity of statistics between these two regions.

(19)

F measures the average squared color error of the segments, penalizing over-segmentation by weighting proportional to the square root of the number of regions. It is user-independent (no parameter should be fixed by the user) and is image-independent (any contents and type of image can be evaluated) [Liu and Yang, 1994].

F uses the sum of the squared color error of each region, averaged by the square root of region size.

F(I) = L

L l=1

err2F(Rl)

|Rl| (50) F was proposed to improve F, because F was found to have a bias towards over-segmentation, which is the characteristic of producing many more regions than desired within a single real-world object. SinceF favors segmentations with a large number of small regions, F extendedF by penalizing segmentations that have many small regions of the same size [Borsotti et al., 1998].

F(I) = 1 103|Sc|

vu

utM axArea

a=1

N R1+a1(a)

L l=1

err2F(Rl)

|Rl| (51) where:N R(a)= number of regions in the segmented imageScwith area=a;

Similar to F, F uses the sum of the squared color error of each region, averaged by the square root of region size.

Qis an improved version ofF since it deals to both over-segmentation and under-segmentation problems [Borsotti et al., 1998].

Q(I) =

√L 103|Sc|

L l=1

[

err2F(Rl) 1 + log(|Rl|) +

(N R(|Rl|)

|Rl|

)2]

(52) Quses the sum of the squared color error of each region averaged by the logarithm of the region size. In order to deal to the very small regions, a constant (equal to 1) is added to this logarithm.

Characteristics:

• Range: not specified

• Optimisation direction: not specified

• Other names: -

• Advantages: not specified

• Disadvantages: not specified 6.1.3 Measures based on texture

Busyness Evaluation method Busy measures the ”busyness” of an image, assuming that a ”smoother” image is preferred. The ”busyness” of an image is computed as either the sum of the absolute values of 4- (or 8-) neighbor Laplacians, or is based on the gray-level co-occurrence matrix of the image. Both metrics actually measure the texture or edges across the whole image, so Busy effectively only measures global texture uni- formity, not individual region uniformity. Busy is based on the measure of ”busyness” in the image, with the assumption that the ideal objects and background are not strongly textured and have simple compact shapes [Weszka and Rosenfeld, 1978].

Characteristics:

• Range: not specified

• Optimisation direction: not specified

Referințe

DOCUMENTE SIMILARE

are considered on the standard contour (unitary circle with the centre in zero or the segrnent of a straight line or the real axis; the case of other contours

effects of English influence on Romanian informal conversations from social media, particularly on linguistic interference at interactional level where we can witness instances

RADU-IOSIF ERDELY, MĂRIOARA POENARU, ALIN HORIA MARIN, CAIUS DOROS, HORATIU EUGEN TEFĂNESCU, NICOLAE- CONSTANTIN BALICA, DELIA HORHAT, MIHAELA PRODEA, OCTAVIA MURARIU,

In contrast to Romania, Ukraine, due to significant volumes of foreign currency in the shadow economy and a constantly deteriorating trade balance, has already reached

Key Words: Buddhism, Confucianism, moral cosmopolitism, Christianism, democracy, universal human rights, ethics, Islam, critical postmodernism, international relations.. George

Faced with the possible insurrection of the body and of the sensible in general, and in order to surpass the possible nefarious consequences of such a logical dead end, (clear

It is a cosmeticizing tool (donations, sponsorships, tree plantings, etc.) and not of planning a long-term development. There are few real CSR programs in Romania, programs

truths, as we call them; 2) within the unceasing aspiration of overcoming the partial truths and asymptotical closeness to “the essence of truth”, the comprehensive truth is