Solid Waste Image Classification Using Deep Convolutional Neural Network
Separating household waste into categories such as <i>organic</i> and <i>recyclable</i> is a critical part of waste management systems to make sure that valuable materials are recycled and utilised. This is beneficial to human health and the environment because less risky tre...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2022-03-01
|
Series: | Infrastructures |
Subjects: | |
Online Access: | https://www.mdpi.com/2412-3811/7/4/47 |
_version_ | 1797445852424306688 |
---|---|
author | Nonso Nnamoko Joseph Barrowclough Jack Procter |
author_facet | Nonso Nnamoko Joseph Barrowclough Jack Procter |
author_sort | Nonso Nnamoko |
collection | DOAJ |
description | Separating household waste into categories such as <i>organic</i> and <i>recyclable</i> is a critical part of waste management systems to make sure that valuable materials are recycled and utilised. This is beneficial to human health and the environment because less risky treatments are used at landfill and/or incineration, ultimately leading to improved circular economy. Conventional waste separation relies heavily on manual separation of objects by humans, which is inefficient, expensive, time consuming, and prone to subjective errors caused by limited knowledge of waste classification. However, advances in artificial intelligence research has led to the adoption of machine learning algorithms to improve the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> of waste classification from images. In this paper, we used a waste classification dataset to evaluate the performance of a bespoke five-layer convolutional neural network when trained with two different image resolutions. The dataset is publicly available and contains 25,077 images categorised into 13,966 organic and 11,111 recyclable waste. Many researchers have used the same dataset to evaluate their proposed methods with varying <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> results. However, these results are not directly comparable to our approach due to fundamental issues observed in their method and validation approach, including the lack of transparency in the experimental setup, which makes it impossible to replicate results. Another common issue associated with image classification is high computational cost which often results to high development time and prediction model size. Therefore, a lightweight model with high <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> and a high level of methodology transparency is of particular importance in this domain. To investigate the computational cost issue, we used two image resolution sizes (i.e., <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>225</mn><mo>×</mo><mn>264</mn></mrow></semantics></math></inline-formula> and <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>80</mn><mo>×</mo><mn>45</mn></mrow></semantics></math></inline-formula>) to explore the performance of our bespoke five-layer convolutional neural network in terms of development time, model size, predictive <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula>, and cross-entropy <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>l</mi><mi>o</mi><mi>s</mi><mi>s</mi></mrow></semantics></math></inline-formula>. Our intuition is that smaller image resolution will lead to a lightweight model with relatively high and/or comparable <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> than the model trained with higher image resolution. In the absence of reliable baseline studies to compare our bespoke convolutional network in terms of <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> and <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>l</mi><mi>o</mi><mi>s</mi><mi>s</mi></mrow></semantics></math></inline-formula>, we trained a random guess classifier to compare our results. The results show that small image resolution leads to a lighter model with less training time and the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> produced (80.88%) is better than the 76.19% yielded by the larger model. Both the small and large models performed better than the baseline which produced 50.05% <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula>. To encourage reproducibility of our results, all the experimental artifacts including preprocessed dataset and source code used in our experiments are made available in a public repository. |
first_indexed | 2024-03-09T13:31:45Z |
format | Article |
id | doaj.art-feb6850a9fc847a5b8c08df2fcd0f9c7 |
institution | Directory Open Access Journal |
issn | 2412-3811 |
language | English |
last_indexed | 2024-03-09T13:31:45Z |
publishDate | 2022-03-01 |
publisher | MDPI AG |
record_format | Article |
series | Infrastructures |
spelling | doaj.art-feb6850a9fc847a5b8c08df2fcd0f9c72023-11-30T21:17:28ZengMDPI AGInfrastructures2412-38112022-03-01744710.3390/infrastructures7040047Solid Waste Image Classification Using Deep Convolutional Neural NetworkNonso Nnamoko0Joseph Barrowclough1Jack Procter2Department of Computer Science, Edge Hill University, St Helens Rd., Ormskirk L39 4QP, UKDepartment of Computer Science, Edge Hill University, St Helens Rd., Ormskirk L39 4QP, UKDepartment of Computer Science, Edge Hill University, St Helens Rd., Ormskirk L39 4QP, UKSeparating household waste into categories such as <i>organic</i> and <i>recyclable</i> is a critical part of waste management systems to make sure that valuable materials are recycled and utilised. This is beneficial to human health and the environment because less risky treatments are used at landfill and/or incineration, ultimately leading to improved circular economy. Conventional waste separation relies heavily on manual separation of objects by humans, which is inefficient, expensive, time consuming, and prone to subjective errors caused by limited knowledge of waste classification. However, advances in artificial intelligence research has led to the adoption of machine learning algorithms to improve the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> of waste classification from images. In this paper, we used a waste classification dataset to evaluate the performance of a bespoke five-layer convolutional neural network when trained with two different image resolutions. The dataset is publicly available and contains 25,077 images categorised into 13,966 organic and 11,111 recyclable waste. Many researchers have used the same dataset to evaluate their proposed methods with varying <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> results. However, these results are not directly comparable to our approach due to fundamental issues observed in their method and validation approach, including the lack of transparency in the experimental setup, which makes it impossible to replicate results. Another common issue associated with image classification is high computational cost which often results to high development time and prediction model size. Therefore, a lightweight model with high <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> and a high level of methodology transparency is of particular importance in this domain. To investigate the computational cost issue, we used two image resolution sizes (i.e., <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>225</mn><mo>×</mo><mn>264</mn></mrow></semantics></math></inline-formula> and <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mn>80</mn><mo>×</mo><mn>45</mn></mrow></semantics></math></inline-formula>) to explore the performance of our bespoke five-layer convolutional neural network in terms of development time, model size, predictive <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula>, and cross-entropy <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>l</mi><mi>o</mi><mi>s</mi><mi>s</mi></mrow></semantics></math></inline-formula>. Our intuition is that smaller image resolution will lead to a lightweight model with relatively high and/or comparable <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> than the model trained with higher image resolution. In the absence of reliable baseline studies to compare our bespoke convolutional network in terms of <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> and <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>l</mi><mi>o</mi><mi>s</mi><mi>s</mi></mrow></semantics></math></inline-formula>, we trained a random guess classifier to compare our results. The results show that small image resolution leads to a lighter model with less training time and the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula> produced (80.88%) is better than the 76.19% yielded by the larger model. Both the small and large models performed better than the baseline which produced 50.05% <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mrow><mi>a</mi><mi>c</mi><mi>c</mi><mi>u</mi><mi>r</mi><mi>a</mi><mi>c</mi><mi>y</mi></mrow></semantics></math></inline-formula>. To encourage reproducibility of our results, all the experimental artifacts including preprocessed dataset and source code used in our experiments are made available in a public repository.https://www.mdpi.com/2412-3811/7/4/47image classificationwaste managementimage processingdeep learningmachine learningwaste recognition |
spellingShingle | Nonso Nnamoko Joseph Barrowclough Jack Procter Solid Waste Image Classification Using Deep Convolutional Neural Network Infrastructures image classification waste management image processing deep learning machine learning waste recognition |
title | Solid Waste Image Classification Using Deep Convolutional Neural Network |
title_full | Solid Waste Image Classification Using Deep Convolutional Neural Network |
title_fullStr | Solid Waste Image Classification Using Deep Convolutional Neural Network |
title_full_unstemmed | Solid Waste Image Classification Using Deep Convolutional Neural Network |
title_short | Solid Waste Image Classification Using Deep Convolutional Neural Network |
title_sort | solid waste image classification using deep convolutional neural network |
topic | image classification waste management image processing deep learning machine learning waste recognition |
url | https://www.mdpi.com/2412-3811/7/4/47 |
work_keys_str_mv | AT nonsonnamoko solidwasteimageclassificationusingdeepconvolutionalneuralnetwork AT josephbarrowclough solidwasteimageclassificationusingdeepconvolutionalneuralnetwork AT jackprocter solidwasteimageclassificationusingdeepconvolutionalneuralnetwork |