Automated waste classification for smart recycling: A multi-class CNN approach with transfer learning and pre-trained models


GÜRCAN F., Soylu A.

ENVIRONMENTAL TECHNOLOGY & INNOVATION, cilt.41, 2026 (SCI-Expanded, Scopus) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 41
  • Basım Tarihi: 2026
  • Doi Numarası: 10.1016/j.eti.2025.104673
  • Dergi Adı: ENVIRONMENTAL TECHNOLOGY & INNOVATION
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus, BIOSIS, Compendex, Directory of Open Access Journals
  • Karadeniz Teknik Üniversitesi Adresli: Evet

Özet

This study presents a comparative analysis of pre-trained CNN models using transfer learning for the multi-class classification of recyclable waste. The dataset used is publicly available and comprises 15,150 images belonging to 12 different categories of recyclable household waste: paper, cardboard, biological, metal, plastic, green-glass, brown-glass, white-glass, clothes, shoes, batteries, and trash. The methodology involves implementing 14 different CNN models pretrained on ImageNet (e.g., ConvNeXt_V1, ResNet50, MobileNet_V3_Large) on this dataset through a transfer learning procedure. In all pre-trained models, only the final classifier layer was trained, while all other layers were kept frozen, utilizing a feature extraction strategy that leverages previously learned visual representations. This approach significantly reduced training time and computational cost, while minimizing the risk of overfitting in scenarios with limited data. Model performance was evaluated using classification metrics such as accuracy, F1-score, precision, and recall, as well as computational efficiency indicators including architectural depth, parameter count, and training time. According to the findings, the ConvNeXt_V1 model achieved the highest performance with a test accuracy of 96.95 %. On the other hand, MobileNet_V2 (93.90 %), MobileNet_V3_Large (93.69 %), MNASNet_1_0 (93.69 %), RegNet_Y_800MF (92.87 %), and SqueezeNet_1_0 (92.10 %) offered shorter training times (ranging from 2856 to 7219 s) and stood out as efficient alternatives for embedded system applications.