Next Article in Journal
Reduced Data Volumes through Hybrid Machine Learning Compared to Conventional Machine Learning Demonstrated on Bearing Fault Classification
Next Article in Special Issue
SiamCAM: A Real-Time Siamese Network for Object Tracking with Compensating Attention Mechanism
Previous Article in Journal
The Effect of Dynamic Lighting for Working Shift People on Clinical Heart Rate Variability and Human Slow Wave Sleep
Previous Article in Special Issue
Modeling Pedestrian Motion in Crowded Scenes Based on the Shortest Path Principle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Small Object Detection and Classification Model for Garbage Waste Management in Smart Cities and IoT Environment

by
Faisal S. Alsubaei
1,
Fahd N. Al-Wesabi
2,3,* and
Anwer Mustafa Hilal
4
1
Department of Cybersecurity, College of Computer Science and Engineering, University of Jeddah, Jeddah 21959, Saudi Arabia
2
Department of Computer Science, College of Science & Art at Mahayil, King Khalid University, Abha 62529, Saudi Arabia
3
Department of Information Systems, College of Computer and Information Technology, Sana’a University, Sana’a 12544, Yemen
4
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam bin Abdulaziz University, AlKharj 16278, Saudi Arabia
*
Author to whom correspondence should be addressed.
Submission received: 12 January 2022 / Revised: 11 February 2022 / Accepted: 15 February 2022 / Published: 22 February 2022
(This article belongs to the Special Issue Deep Learning in Object Detection and Tracking)

Abstract

:
In recent years, object detection has gained significant interest and is considered a challenging problem in computer vision. Object detection is mainly employed for several applications, such as instance segmentation, object tracking, image captioning, healthcare, etc. Recent studies have reported that deep learning (DL) models can be employed for effective object detection compared to traditional methods. The rapid urbanization of smart cities necessitates the design of intelligent and automated waste management techniques for effective recycling of waste. In this view, this study develops a novel deep learning-based small object detection and classification model for garbage waste management (DLSODC-GWM) technique. The proposed DLSODC-GWM technique mainly focuses on detecting and classifying small garbage waste objects to assist intelligent waste management systems. The DLSODC-GWM technique follows two major processes, namely, object detection and classification. For object detection, an arithmetic optimization algorithm (AOA) with an improved RefineDet (IRD) model is applied, where the hyperparameters of the IRD model are optimally chosen by the AOA. Secondly, the functional link neural network (FLNN) technique was applied for the classification of waste objects into multiple classes. The design of IRD for waste classification and AOA-based hyperparameter tuning demonstrates the novelty of the work. The performance validation of the DLSODC-GWM technique is performed using benchmark datasets, and the experimental results show the promising performance of the DLSODC-GWM method on existing approaches with a maximum a c c u y of 98.61%.

1. Introduction

With the increase in smart video surveillance, facial detection, autonomous vehicles, and plenty of people counting applications, accurate and fast object detection methods are in increasing demand. Such a system involves classifying and recognizing all the objects in the image but localizing each one by drawing the proper bounding box around it [1]. This makes object detection a considerably more difficult process than its conventional computer vision (CV) predecessor, image classification. The current progression in the fields of deep learning (DL), image processing, and CV technologies has changed the thinking about different features of day-to-day lives [2]. The DL method has given a strong foundation for image detection with consistent accuracy [3]. The widespread image classification, the convolution neural networks (CNNs), is stimulated by biological neural networks that are comprised of different neurons and layers in every layer are closely linked to the neurons in the following layer [4]. The advantages of utilizing a CNN that provides independence between prior knowledge, the least amount of effort in design, and feature extraction. CNNs have made great achievements in image classification and recognition [5]. The accuracy and popularity of CNNs for the classification of images have improved because of the large-scale integrated systems for image processing and learning, the vast accessibility of public data of images, and higher-speed GPUs. The concept of smart waste classification using trash and waste images has great potential.
Because of fast urbanization, nowadays, cities are facing several problems. One of them is waste management systems, since the amount of waste is directly proportionate to the number of people who live in urban areas. The city administrations and the municipalities utilize conventional waste classification methods that are very slow, manual, costly, and inefficient [6]. Consequently, automated waste management and classification are indispensable for the city that is being urbanized for the improved recycling of waste. The simplification of the waste classification method is needed since the technologies have been rapidly growing and many manual tasks have been decreased by adopting artificial intelligence (AI) methods [7]. Generally, in the Indian context, waste consists of plastic, paper, metal, rubber, textiles, glass, sanitary products, organics, electronics, and electrical, infectious materials (hospital and clinical), and hazardous substances (paint, spray, and chemicals) [8], are widely categorized as biodegradable (BD) and non-biodegradable (NBD) waste, with a corresponding share of 52% and 48% [9].
Effective waste segregation could help in the appropriate recycling and disposal of this waste according to biodegradability. Therefore, the current era dictates the progress of a smart waste segregation method to suggest the abovementioned cause of ecological damage. Consequently, the segregation of waste has received considerable interest from academicians and researchers worldwide [10]. The proper organization and classification of waste into different classes (such as biodegradable, recyclable, organic, harmful, non-biodegradable, and so on.) assists in the appropriate disposal and utilization of waste. For waste segregation, the CV technique might offer an efficient solution to separate out, identify, and classify the waste from the huge dumps of trash and garbage.
Kumar et al. [11] examine a method for waste segregation for its efficient disposal and recycling through a DL approach. The YOLOv3 algorithm was used from the Darknet neural network for training self-made datasets. The network was trained for six object types (such as glass, cardboard, paper, metal, organic, and plastic waste). Furthermore, for relative analysis, the detection process was implemented by utilizing YOLOv3-tiny for validating the capability of the YOLOv3. Nasrullah et al. [12] employed two deep 3D customized mixed link network (CMixNet) frameworks for lung nodule classification and recognition, correspondingly. Nodule classifications were implemented by using a GBM system on the learned features from the 3D CMixNet framework. Nodule detection was implemented by using fast RCNN on effectively-learned features from U-Net and CMixNet, such as the encoder–decoder framework. Hiary et al. [13] presented two-phase DL classifiers to differentiate flowers of a wide-ranging species. Firstly, the flower region is segmented automatically to permit localization of the minimal bounding box around it. The presented method is modeled as a binary classification in a full convolution network architecture. Next, a strong CNN classification is constructed to differentiate the variety of flowers.
Chen et al. [14] developed three post-processing methods to increase the benchmark Fast RCNN based on previous knowledge. Firstly, a filtering method is created to remove over-lapping boxes identified by Fast RCNN related to a similar tooth. Then, a NN system is performed to detect missing teeth. At last, a rule-based model based on a teeth number scheme is projected for matching labels of identified teeth boxes to alter outcomes that violate some intuitive rules. Vo et al. [15] presented a strong method using DNN for automatically classifying trash that is employed in smart waste sorter machines. First, it collects the VN-trash data set, which comprises 5904 images belonging to the following three distinct groups: medical, organic, and inorganic wastes from Vietnam. Then, we developed a DNN system for trash classification called DNN-TC, which is an enhancement of ResNeXt to optimize the performance of prediction.
Ahmad et al. [16] introduced a method called “double fusion” that optimally integrates various DL methods using score-level fusion and feature models. The double fusion system guarantees an enhanced contribution of the deep model by, initially, integrating the capability of earlier and latter fusion systems and a score-level fusion of the classification result attained by earlier and latter fusion models. Sheng et al. [17] designed a smart waste management method with LoRa transmission and a TensorFlow-based DL method. The presented method transmits the sensor data and Tensorflow implements real-time object classification and detection. The bin contains a number of chambers for segregating the waste involving plastic, metal, paper, and common waste compartments that are managed by the servo motor.
This study develops a novel deep learning-based small object detection and classification model for garbage waste management (DLSODC-GWM) technique. The proposed DLSODC-GWM involves the design of an arithmetic optimization algorithm (AOA) with an improved RefineDet (IRD) model for an effectual object detection process where the hyperparameters of the IRD model are optimally chosen by the AOA. In addition, the functional link neural network (FLNN) model is applied for the classification of waste objects into multiple classes. In order to demonstrate the significant performance of the DLSODC-GWM approach, a wide-ranging simulation analysis is carried out on benchmark datasets.

2. The Proposed DLSODC-GWM Technique

In this study, a new DLSODC-GWM technique has been developed for waste management systems in order to effectually detect and classify small garbage waste objects. The DLSODC-GWM technique involves three distinct subprocesses, namely, IRD object recognition, hyperparameter tuning, and FLNN-based object classification. During the object detection process, the AOA is applied to optimally select the hyperparameter values of the IRD model and thereby improve the detection efficiency. The detailed workings of these three modules are elaborated on in the following.

2.1. IRD-Based Object Detection Module

The improved RefineDet method using VGG16 as the central network [18] creates a sequence of anchors with distinct aspect ratios and scales in all the feature maps by utilizing the anchor generation method of RPN, and also attains a fixed number of objects bounding boxes afterward two regressions and classifications, the possibility of the occurrence of distinct groups in this bounding box. Eventually, the last regression and classification outcomes are attained by using non-maximal suppression (NMS). The enhanced RefineDet method is classified into the object detection module (ODM), the transfer connection block (TCB), and the anchor refinement module (ARM). A certain network framework has been demonstrated in Figure 1.

2.1.1. ARM Module

It is largely comprised of the VGG16 and the convolutional layer. The ARM implements anchor generation, anchor refinement, negative anchor filtering, and feature extraction. Once the confidence of negative samples is superior when compared to 0.99, the module rejects and does not utilize them for the concluding detection in ODM. During the feature extraction process, two convolutional layers, that is, conv6_1 and conv6_2, are added to the VGG16 network. The negative anchor filtering efficiently filters out the negative anchor box using well-classification; furthermore, this improves the sample imbalances. Next, add the following four other convolutional layers: conv7_1, conv7_2, conv8_1, and conv8_2 to capture higher-level semantic data. In addition, the higher-level feature of conv8_2 is merged with the lower-level feature of conv7_2. Next, the combined feature is transmitted to the low-level feature through TCB, thus the low-level feature map is utilized for detecting high semantic data and improving the recognition performance of the floating object.

2.1.2. TCB Module

It is commonly utilized for connecting ODM and ARM and transferring the feature data of ARM to ODM. Additionally, akin to the framework of FPN, adjacent TCB is interconnected for realizing the feature of higher- and lower-level features and increases the semantic data of lower-level features.

2.1.3. ODM Module

It has largely consisted of the output of TCB and the predictive layers (regression and classification layers, that is., the convolutional layer with a 3 × 3 kernel size). In addition, the output of the predictive layer is the certain class of the refined anchor and the coordinate offset in relation to the refined anchor boxes. The refined anchor is utilized as input for regression and classification, and the last bounding boxes are chosen on the basis of NMS.

2.1.4. Loss Function

The IRD technique utilizes anchors that are intensively sampled on distinct scale feature maps as trained samples. Most of its samples are easy negative samples that do not comprise the object, and only some positive instances that comprise the object. As a result, the imbalance of positive as well as negative samples are magnified. Numerous easy negative samples have minimum promotion effects on the concept from the entire trained procedure and simulate the final detection efficiency of the method. The focal loss function (LF) was dynamic cross-entropy (CE) LFs. According to the novel function, the CE-LF was redesigned. The weight factor α and modulation factors ( 1 p t ) y are established, in which the α factor balances positive as well as negative samples, and the modulation factor ( 1 p t ) γ modifies the weight of easy and hard samples to make the method effort on hard examples under the trained procedure and enhance the accuracy. The focal LFs are as follows:
F L ( p t ) = α t ( 1 p t ) γ   log   ( p t ) ,
where,
p t = { p     i f   y = 1 1 p o t h e r w i s e ,    
where y refers the ground truth class labels containing if the sample was foreground or not. When it can be foreground, the label has 1, else it can be −1. Besides, p [ O , 1 ] defines the type probabilities of foreground samples with label y = 1 , for instance, the forecast possibility of a sample comprising the objects.
α t = { α     i f   y = 1 1 α o t h e r w i s e .    
While the weight factor, α weakens the result of the easy sample loss value on the entire network loss and balances the effect of positive, as well as negative samples, on the networks.

2.2. AOA-Based Hyperparameter Tuning Module

Usually, as another MH technique, the AOA contains two search stages [19], namely, exploration and exploitation, which are simulated as mathematics functions such as , + ,     , and / . Initial, the AOA makes a group of N solution (agent). All but one demonstrates the agent to the tested problems. Therefore, the solution/agent signifies the X population, as follows:
X = [ x N 1 , 1 x N , 1 x 2 , 1 x 1 , 1   x N 1 , j x N , j x 2 , j x 1 , j   x N , n 1 x 1 , n 1   x N 1 , n x N , n x 2 , n x 1 , n ]
X = [ x 1 , 1 x 1 , j x 1 , n 1 x 1 , n x 2 , 1 x 2 , j x 2 , n x N 1 , 1 x N 1 , j x N 1 , n x N , 1 x N , j x N , n 1 x N , n ]  
Afterward, the FF of all solutions is calculated to detect the optimum one X b . Next, based on the Math Optimizer Accelerated (MOA) value, the AOA executes exploration or exploitation procedures. At that time, M O A was upgraded as the following subsequent formula:
M O A ( t ) =   Min   + t × ( Max M O A Min M O A M t )  
where M t denotes the entire number of iterations. Min M O A and signifies the minimal and maximal values of accelerated functions, correspondingly. In particular, multiplication (M) and division (D) were used from the exploration stage of AOA as obtainable from the following subsequent formula:
X i , j ( t + 1 ) = { X b j ÷ ( M O P + ) × ( ( U B j L B j ) × μ + L B j ) ,     r 2 < 0.5 X b j × M O P × ( ( U B j L B j ) × μ + L B j ) , o t h e r w i s e
where e implies the smaller integer values, U B j and L B j stands for the lower and upper boundaries of the searching region at j t h dimensional. μ = 0.5 refers the control functions. In addition, the Math Optimizer ( M O P ) was explained as follows:
M O P ( t ) = 1 t 1 / α M t 1 / α
α = 5 defines the dynamic parameter that resolves the precision of the exploitation stage throughout iterations.
Moreover, addition (A) and subtracting (D) functions were utilized for implementing the AOA exploitation stage utilizing the subsequent formula.
x i , j ( t + 1 ) = { X b j M O P × ( ( U B j L B j ) × μ + L B j ) ,   r 3 < 0.5 X b j + M O P × ( ( U B j L B j ) × μ + L B j ) ,   o t h e r w i s e
where r 3 implies the arbitrary number generated between zero and one. Then, the solution’s upgrade method was executed utilizing the AOA operator. In summary, Algorithm 1 defines the important stages of the AOA.
Algorithm 1 Steps involved in AOA
Input: The parameter of AOA-like dynamic exploitation parameters (α), control function amount of solutions (N), and entire amount of iterations Mt·
Create the primary value to solutions Xii = 1, … N
while (t < Mt) do
  Calculate the FF to all the solutions.
  Define the optimum solutions Xb
  Upgrade the MOA and MOP utilizing in Equations (3) and (5) correspondingly
  for i = 1 to N do
    for j = 1 to Dim do
      Upgrade the value of r1, r2 and r3.
      if r1 > MOA then
        Exploration stage
        Utilize Equation (4) for updating the Xi
      else
        Exploitation stage
        Utilize Equation (6) for updating the Xi
      end if
    end for
  end for
            t = t + 1
end while
Output the optimum solution (feature subset) (Xb)

2.3. FLNN-Based Object Classification Module

Next to the object detection process, the classification module using the FLNN technique is executed to allot distinct class labels. The FLNN is the part of FFNN without a hidden layer [20]. It takes non-linearity to their input design by extension unit. This unit assists as more input to networks and the sum of these units are calculated at the resultant layer. The extensive input to the network decreased the calculation cost and other sides, it can be enhanced by the calculation efficiency as connected to BP. In the meantime, the inherited property of the invariant effort the FLANN for picking up only the chosen signals that obtained optimum system identification. The second order FLNN infrastructure that is collected of three inputs, x 1 , x 2 , x 3 and with around higher-order combination.
Y i = σ ( W o + j w j x j + j , k w j k x j x k + j , k , l w j k l x j x k x l )
W o implies the tunable threshold and σ signifies the non-linear transfer functions.

3. Performance Validation

The proposed DLSODC-GWM technique is simulated using the Python 3.6.5 tool. The proposed model is tested using a PC MSI Z370-A Pro with an i5-8600k processor, GeForce GTX 1050 Ti 4 GB, 16 GB of RAM, a 250 GB SSD, and a 1 TB HDD. For experimental validation, a ten-fold cross-validation process is employed. The parameter setting is given as follows: batch size: 128, learning rate: 0.001, momentum: 0.2, and optimizer: AOA. The experimental result analysis of the DLSODC-GWM technique is examined using the benchmark garbage classification dataset from the Kaggle repository [21]. The dataset contains 393 images under the cardboard class, 491 images under the glass class, 400 images under the metal class, 584 images under the paper class, 472 images under the plastic class, and 127 images under the trash class. A few sample images are illustrated in Figure 2.
The heat map analysis of the objects that exist in the dataset is shown in Figure 3. Figure 4 visualizes the sample object detection outcomes of the DLSODC-GWM technique on the test images applied. From the figure, it is obvious that the DLSODC-GWM technique has identified glass, metal, and trash objects with the maximum a c c u y of 99%.
The confusion matrix generated by the DLSODC-GWM technique on the classification of waste under 1000 epochs is in Figure 5. The figure reported that the DLSODC-GWM technique has identified 372 images into cardboard, 472 images into glass, 380 images into metal, 570 images into paper, 463 images into plastic, and 107 images into trash.
The classifier results of the DLSODC-GWM technique on the classification of waste objects with 1000 epochs are given in Table 1 and Figure 6. The table values pointed out that the DLSODC-GWM technique has effectually recognized all the class labels. For instance, the DLSODC-GWM technique has categorized into images into cardboard with p r e c n , r e c a l , a c c u y , and F s c o r e of 95.88%, 94.66%, 98.50%, and 95.26%, respectively. On the other hand, the DLSODC-GWM technique has recognized images into paper with p r e c n , r e c a l , a c c u y , and F s c o r e of 96.77%, 97.60%, 98.66%, and 97.19%, respectively.
The accuracy outcome analysis of the DLSODC-GWM technique on the test data is portrayed in Figure 7. The results demonstrated that the DLSODC-GWM technique has accomplished improved validation accuracy compared to training accuracy. It is also observable that the accuracy values become saturated with the epoch count of 1000.
The loss outcome analysis of the DLSODC-GWM technique on the test data is depicted in Figure 8. The figure reveals that the DLSODC-GWM technique has denoted the reduced validation loss over the training loss. It is additionally noticed that the loss values become saturated with the epoch count of 1000.
The confusion matrix generated by the DLSODC-GWM approach to the classification of waste under 2000 epochs in Figure 9. The figure described that the DLSODC-GWM methodology has identified 372 images into cardboard, 467 images into glass, 373 images into metal, 565 images into paper, 456 images into plastic, and 114 images into trash.
The classifier outcomes of the DLSODC-GWM algorithm on the classification of waste objects with 2000 epochs are provided in Table 2 and Figure 10. The table values pointed out that the DLSODC-GWM system has effectually recognized all the class labels. For instance, the DLSODC-GWM method has categorized images into cardboard with p r e c n , r e c a l , a c c u y , and F s c o r e of 95.38%, 94.66%, 98.42%, and 95.02%, correspondingly. In addition, the DLSODC-GWM approach has recognized images into paper with p r e c n , r e c a l , a c c u y , and F s c o r e of 95.28%, 96.75%, 98.09%, and 96.01%, correspondingly.
The accuracy outcome analysis of the DLSODC-GWM technique on the test data is illustrated in Figure 11. The outcomes exhibited that the DLSODC-GWM approach has accomplished improved validation accuracy compared to training accuracy. It can also be observed that the accuracy values become saturated with the epoch count of 1000.
The loss outcome analysis of the DLSODC-GWM technique on the test data is outperformed in Figure 12. The figure shows that the DLSODC-GWM technique has denoted the lower validation loss over the training loss. It is additionally observed that the loss values become saturated with the epoch count of 1000.
Table 3 and Figure 13 demonstrate the comparative a c c u y analysis of the DLSODC-GWM technique with recent methods. The results show that the AlexNet model has resulted in lower performance with a c c u y of 52.50%. In line with this, the VGG16 model has obtained a greatly improved a c c u y of 73.10%; whereas, the ResNet50 model has accomplished and even greater increased a c c u y of 74.70%. Though the MLH-CNN technique has resulted in a near-optimal a c c u y of 92.60%, the presented DLSODC-GWM technique has accomplished superior performance with a c c u y of 98.61%.
Table 4 and Figure 14 showcase the comparative p r e c n , r e c a l , and F s c o r e analysis of the DLSODC-GWM method with recent algorithms [22]. The outcomes demonstrated that the AlexNet approach has resulted in lower performance with p r e c n , r e c a l , and F s c o r e of 42%, 50%, and 44%, correspondingly. Followed by, the VGG16 system has achieved a greatly improved p r e c n , r e c a l , and F s c o r e of 69%, 68%, and 68%, respectively; whereas, the ResNet50 model has accomplished even increased p r e c n , r e c a l , and F s c o r e of 72%, 72%, and 72%, respectively.
However, the MLH-CNN approach has resulted in near-optimal p r e c n , r e c a l , and F s c o r e of 91%, 91%, and 91%, respectively. The presented DLSODC-GWM methodology has accomplished superior performance with p r e c n , r e c a l , and F s c o r e of 95.23%, 94.29%, and 94.73%, respectively.
From the aforementioned results and discussion, it can be stated that the DLSODC-GWM technique has accomplished enhanced waste object classification performance compared to existing techniques.

4. Conclusions

In this study, a new DLSODC-GWM technique has been developed for waste management systems in order to effectually detect and classify small garbage waste objects. The DLSODC-GWM technique involves three distinct subprocesses, namely, IRD object recognition, hyperparameter tuning, and FLNN-based object classification. During the object detection process, the AOA is applied to optimally select the hyperparameter values of the IRD model and thereby improve the detection efficiency. For demonstrating the significant performance of the DLSODC-GWM technique, a wide-ranging simulation analysis is carried out on benchmark datasets. The extensive comparative analysis highlighted the superior outcomes of the DLSODC-GWM approach over existing approaches. Therefore, the DLSODC-GWM technique has the ability to proficiently identify and classify small objects in the waste management system. In the future, fusion of DL models can be employed to enhance the detection efficiency of the DLSODC-GWM technique.

Author Contributions

Conceptualization, F.S.A. and A.M.H.; methodology, F.N.A.-W.; software, A.M.H.; validation, F.S.A., F.N.A.-W. and A.M.H.; formal analysis, F.N.A.-W.; investigation, A.M.H.; resources, F.S.A.; data curation, F.N.A.-W.; writing—original draft preparation, F.S.A.; writing—review and editing, A.M.H.; visualization, F.N.A.-W.; supervision, F.S.A.; project administration, F.N.A.-W.; funding acquisition, F.S.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number (MoE-IF-20-02/10).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all participants involved in the study.

Data Availability Statement

Data is available on reasonable request.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this research work through the project number (MoE-IF-20-02/10).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, X.; Sahoo, D.; Hoi, S.C. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef] [Green Version]
  2. Jiang, Q.; Tan, D.; Li, Y.; Ji, S.; Cai, C.; Zheng, Q. Object detection and classification of metal polishing shaft surface defects based on convolutional neural network deep learning. Appl. Sci. 2020, 10, 87. [Google Scholar] [CrossRef] [Green Version]
  3. Vaidya, B.; Paunwala, C. Deep learning architectures for object detection and classification. In Smart Techniques for a Smarter Planet; Springer: Cham, Switzerland, 2019; pp. 53–79. [Google Scholar]
  4. Pal, S.K.; Pramanik, A.; Maiti, J.; Mitra, P. Deep learning in multi-object detection and tracking: State of the art. Appl. Intell. 2021, 51, 6400–6429. [Google Scholar] [CrossRef]
  5. Zheng, Y.Y.; Kong, J.L.; Jin, X.B.; Wang, X.Y.; Su, T.L.; Zuo, M. CropDeep: The crop vision dataset for deep-learning-based classification and detection in precision agriculture. Sensors 2019, 19, 1058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Adedeji, O.; Wang, Z. Intelligent Waste Classification System Using Deep Learning Convolutional Neural Network. Procedia Manuf. 2019, 35, 607–612. [Google Scholar] [CrossRef]
  7. Melikoglu, M. Reutilisation of food wastes for generating fuels and value added products: A global review. Environ. Technol. Innov. 2020, 19, 101040. [Google Scholar] [CrossRef]
  8. Chu, Y.; Huang, C.; Xie, X.; Tan, B.; Kamal, S.; Xiong, X. Multilayer Hybrid Deep-Learning Method for Waste Classification and Recycling. Comput. Intell. Neurosci. 2018, 2018, 5060857. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Youme, O.; Bayet, T.; Dembele, J.M.; Cambier, C. Deep Learning and Remote Sensing: Detection of Dumping Waste Using UAV. Procedia Comput. Sci. 2021, 185, 361–369. [Google Scholar] [CrossRef]
  10. Li, J.; Chen, J.; Sheng, B.; Li, P.; Yang, P.; Feng, D.D.; Qi, J. Automatic Detection and Classification System of Domestic Waste via Multi-model Cascaded Convolutional Neural Network. IEEE Trans. Ind. Inform. 2022, 18, 163–173. [Google Scholar] [CrossRef]
  11. Kumar, S.; Yadav, D.; Gupta, H.; Verma, O.P.; Ansari, I.A.; Ahn, C.W. A novel yolov3 algorithm-based deep learning approach for waste segregation: Towards smart waste management. Electronics 2021, 10, 14. [Google Scholar] [CrossRef]
  12. Nasrullah, N.; Sang, J.; Alam, M.S.; Mateen, M.; Cai, B.; Hu, H. Automated lung nodule detection and classification using deep learning combined with multiple strategies. Sensors 2019, 19, 3722. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Hiary, H.; Saadeh, H.; Saadeh, M.; Yaqub, M. Flower classification using deep convolutional neural networks. IET Comput. Vis. 2018, 12, 855–862. [Google Scholar] [CrossRef]
  14. Chen, H.; Zhang, K.; Lyu, P.; Li, H.; Zhang, L.; Wu, J.; Lee, C.H. A deep learning approach to automatic teeth detection and numbering based on object detection in dental periapical films. Sci. Rep. 2019, 9, 3840. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Vo, A.H.; Vo, M.T.; Le, T. A novel framework for trash classification using deep transfer learning. IEEE Access 2019, 7, 178631–178639. [Google Scholar] [CrossRef]
  16. Ahmad, K.; Khan, K.; Al-Fuqaha, A. Intelligent fusion of deep features for improved waste classification. IEEE Access 2020, 8, 96495–96504. [Google Scholar] [CrossRef]
  17. Sheng, T.J.; Islam, M.S.; Misran, N.; Baharuddin, M.H.; Arshad, H.; Islam, M.R.; Chowdhury, M.E.; Rmili, H.; Islam, M.T. An internet of things based smart waste management system using LoRa and tensorflow deep learning model. IEEE Access 2020, 8, 148793–148811. [Google Scholar] [CrossRef]
  18. Zhang, L.; Wei, Y.; Wang, H.; Shao, Y.; Shen, J. Real-Time Detection of River Surface Floating Object Based on Improved RefineDet. IEEE Access 2021, 9, 81147–81160. [Google Scholar] [CrossRef]
  19. Abualigah, L.; Diabat, A.; Mirjalili, S.; Abd Elaziz, M.; Gandomi, A.H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  20. Khan, A.; Bukhari, J.; Bangash, J.I.; Khan, A.; Imran, M.; Asim, M.; Ishaq, M.; Khan, A. Optimizing connection weights of functional link neural network using APSO algorithm for medical data classification. J. King Saud Univ. Comput. Inf. Sci. 2020; in press. [Google Scholar]
  21. Kaggle. Garbage Classifiction. Available online: https://www.kaggle.com/asdasdasasdas/garbage-classification (accessed on 12 January 2022).
  22. Shi, C.; Tan, C.; Wang, T.; Wang, L. A waste classification method based on a multilayer hybrid convolution neural network. Appl. Sci. 2021, 11, 8572. [Google Scholar] [CrossRef]
Figure 1. Schematic representation of RefineDet architecture.
Figure 1. Schematic representation of RefineDet architecture.
Applsci 12 02281 g001
Figure 2. Sample images: (a) cardboard, (b) glass, (c) metal, and (d) paper.
Figure 2. Sample images: (a) cardboard, (b) glass, (c) metal, and (d) paper.
Applsci 12 02281 g002
Figure 3. Heat map analysis. (a) Cardboard, (bd) glass.
Figure 3. Heat map analysis. (a) Cardboard, (bd) glass.
Applsci 12 02281 g003
Figure 4. Visualization of sample object detection. (a,b) Glass, (c,d) metal, and (e,f) trash.
Figure 4. Visualization of sample object detection. (a,b) Glass, (c,d) metal, and (e,f) trash.
Applsci 12 02281 g004
Figure 5. Confusion matrix of DLSODC-GWM technique under 1000 epochs.
Figure 5. Confusion matrix of DLSODC-GWM technique under 1000 epochs.
Applsci 12 02281 g005
Figure 6. Result analysis of DLSODC-GWM technique under 1000 epochs.
Figure 6. Result analysis of DLSODC-GWM technique under 1000 epochs.
Applsci 12 02281 g006
Figure 7. Accuracy graph analysis of DLSODC-GWM technique under 1000 epochs.
Figure 7. Accuracy graph analysis of DLSODC-GWM technique under 1000 epochs.
Applsci 12 02281 g007
Figure 8. Loss graph analysis of DLSODC-GWM technique under 1000 epochs.
Figure 8. Loss graph analysis of DLSODC-GWM technique under 1000 epochs.
Applsci 12 02281 g008
Figure 9. Confusion matrix of DLSODC-GWM technique under 2000 epochs.
Figure 9. Confusion matrix of DLSODC-GWM technique under 2000 epochs.
Applsci 12 02281 g009
Figure 10. Result analysis of DLSODC-GWM technique under 2000 epochs.
Figure 10. Result analysis of DLSODC-GWM technique under 2000 epochs.
Applsci 12 02281 g010
Figure 11. Accuracy graph analysis of DLSODC-GWM technique under 1000 epochs.
Figure 11. Accuracy graph analysis of DLSODC-GWM technique under 1000 epochs.
Applsci 12 02281 g011
Figure 12. Loss graph analysis of DLSODC-GWM technique under 1000 epochs.
Figure 12. Loss graph analysis of DLSODC-GWM technique under 1000 epochs.
Applsci 12 02281 g012
Figure 13. Accuracy analysis of DLSODC-GWM technique with recent algorithms.
Figure 13. Accuracy analysis of DLSODC-GWM technique with recent algorithms.
Applsci 12 02281 g013
Figure 14. Comparative analysis of DLSODC-GWM technique with recent methods.
Figure 14. Comparative analysis of DLSODC-GWM technique with recent methods.
Applsci 12 02281 g014
Table 1. Result analysis of DLSODC-GWM technique with different methods under 1000 epochs.
Table 1. Result analysis of DLSODC-GWM technique with different methods under 1000 epochs.
Epoch-1000
MethodsPrecisionRecallAccuracyF-Score
Cardboard95.8894.6698.5095.26
Glass95.9396.1398.4296.03
Metal95.4895.0098.4695.24
Paper96.7797.6098.6697.19
Plastic95.8698.0998.8296.96
Trash91.4584.2598.7887.70
Average95.2394.2998.6194.73
Table 2. Result analysis of DLSODC-GWM technique with different methods under 2000 epochs.
Table 2. Result analysis of DLSODC-GWM technique with different methods under 2000 epochs.
Epoch-2000
MethodsPrecisionRecallAccuracyF-Score
Cardboard95.3894.6698.4295.02
Glass95.8995.1198.2295.50
Metal96.1393.2598.3094.67
Paper95.2896.7598.0996.01
Plastic97.4496.6198.8797.02
Trash80.8589.7698.3885.07
Average93.5094.3698.3893.88
Table 3. Accuracy analysis of DLSODC-GWM technique with recent methods.
Table 3. Accuracy analysis of DLSODC-GWM technique with recent methods.
MethodsAccuracy (%)
DLSODC-GWM 98.61
MLH-CNN92.60
AlexNet52.50
ResNet5074.70
VGG1673.10
Table 4. Comparative analysis of DLSODC-GWM technique with recent approaches.
Table 4. Comparative analysis of DLSODC-GWM technique with recent approaches.
MethodsPrecisionRecallF-Score
DLSODC-GWM 95.2394.2994.73
MLH-CNN91.0091.0091.00
AlexNet42.0050.0044.00
RestNet5072.0072.0072.00
VGG1669.0068.0068.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alsubaei, F.S.; Al-Wesabi, F.N.; Hilal, A.M. Deep Learning-Based Small Object Detection and Classification Model for Garbage Waste Management in Smart Cities and IoT Environment. Appl. Sci. 2022, 12, 2281. https://0-doi-org.brum.beds.ac.uk/10.3390/app12052281

AMA Style

Alsubaei FS, Al-Wesabi FN, Hilal AM. Deep Learning-Based Small Object Detection and Classification Model for Garbage Waste Management in Smart Cities and IoT Environment. Applied Sciences. 2022; 12(5):2281. https://0-doi-org.brum.beds.ac.uk/10.3390/app12052281

Chicago/Turabian Style

Alsubaei, Faisal S., Fahd N. Al-Wesabi, and Anwer Mustafa Hilal. 2022. "Deep Learning-Based Small Object Detection and Classification Model for Garbage Waste Management in Smart Cities and IoT Environment" Applied Sciences 12, no. 5: 2281. https://0-doi-org.brum.beds.ac.uk/10.3390/app12052281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop