A New Performance Evaluation Metric for Classifiers: Polygon Area Metric


Aydemir Ö.

Journal of Classification, cilt.38, ss.16-26, 2021 (SCI-Expanded) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 38
  • Basım Tarihi: 2021
  • Doi Numarası: 10.1007/s00357-020-09362-5
  • Dergi Adı: Journal of Classification
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Social Sciences Citation Index (SSCI), Scopus, Academic Search Premier, BIOSIS, Information Science and Technology Abstracts, INSPEC, Library and Information Science Abstracts, Library, Information Science & Technology Abstracts (LISTA), zbMATH, DIALNET
  • Sayfa Sayıları: ss.16-26
  • Anahtar Kelimeler: Classifier performance, Classification accuracy, Assessment metric, Polygon area metric
  • Karadeniz Teknik Üniversitesi Adresli: Evet

Özet

Classifier performance assessment (CPA) is a challenging task for pattern recognition. In recent years, various CPA metrics have been developed to help assess the performance of classifiers. Although the classification accuracy (CA), which is the most popular metric in pattern recognition area, works well if the classes have equal number of samples, it fails to evaluate the recognition performance of each class when the classes have different number of samples. To overcome this problem, researchers have developed various metrics including sensitivity, specificity, area under curve, Jaccard index, Kappa, and F-measure except CA. Giving many evaluation metrics for assessing the performance of classifiers make large tables possible. Additionally, when comparing classifiers with each other, while a classifier might be more successful on a metric, it may have poor performance for the other metrics. Hence, such kinds of situations make it difficult to track results and compare classifiers. This study proposes a stable and profound knowledge criterion that allows the performance of a classifier to be evaluated with only a single metric called as polygon area metric (PAM). Thus, classifier performance can be easily evaluated without the need for several metrics. The stability and validity of the proposed metric were tested with the k-nearest neighbor, support vector machines, and linear discriminant analysis classifiers on a total of 7 different datasets, five of which were artificial. The results indicate that the proposed PAM method is simple but effective for evaluating classifier performance.