p d the entropy has the 322 partitional clustering

Info iconThis preview shows page 1. Sign up to view the full content.

View Full Document Right Arrow Icon
This is the end of the preview. Sign up to access the rest of the document.

Unformatted text preview: in the literature for inverse purity is microaveraged precision. The reader may note that, in the evaluation of clustering results, microaveraged precision is identical to microaveraged recall (cf. e.g. Sebastiani (2002)). The Fmeasure works similar as inverse purity, but it depreciates overly large clusters, as it includes the individual precision of these clusters into the evaluation. While (inverse) purity and F-measure only consider ‘best’ matches between ‘queries’ and manually defined categories, the entropy indicates how large the Band 20 – 2005 39 Hotho, Nürnberger, and Paaß information content uncertainty of a clustering result with respect to the given classification is E (P , L ) = E( P) = − ∑ prob( P) · E( P), where (15) prob( L| P) log(prob( L| P)) (16) P ∈P ∑ L ∈L where prob( L| P) = Precision( P, L) and prob( P) = range [0, log(|L |)], with 0 indicating optimality. | P| . |D| The entropy has the 3.2.2 Partitional Clustering Manning & Schütze (2001); Steinbach et al. (2000) go...
View Full Document

{[ snackBarMessage ]}

Ask a homework question - tutors are online