International Journal of Fuzzy Logic and Intelligent Systems 2022; 22(3): 223-232
Published online September 25, 2022
https://doi.org/10.5391/IJFIS.2022.22.3.223
© The Korean Institute of Intelligent Systems
Dheo Prasetyo Nugroho, Sigit Widiyanto, and Dini Tri Wardani
Department of Information System Management, Gunadarma University, Depok, Indonesia
Correspondence to :
Dheo Prasetyo Nugroho (dheoprasetyo.dp@gmail.com)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Examination of the technological development in agriculture reveals that not many applications use cameras to detect tomato ripeness; therefore, tomato maturity is still determined manually. Currently, technological advances and developments are occurring rapidly, and are, therefore, also inseparable from the agricultural sector. Object detection can help determining tomato ripeness. In this research, faster region-based convolutional neural network (Faster R-CNN), single shot multibox detector (SSD), and you only look once (YOLO) models were tested to recognize or detect tomato ripeness using input images. The model training process required 5 hours and produced a total loss value <0.5, and as the total loss became smaller, the predicted results improved. Tests were conducted on a training dataset, and average accuracy values of 99.55%, 89.3%, and 94.6% were achieved using the Faster R-CNN, SSD, and YOLO models, respectively.
Keywords: SSD, Faster R-CNN, YOLO, Object detection
No potential conflict of interest relevant to this article was reported.
E-mail: dheoprasetyo.dp@gmail.com
E-mail: dinitri@staff.gunadarma.ac.id
International Journal of Fuzzy Logic and Intelligent Systems 2022; 22(3): 223-232
Published online September 25, 2022 https://doi.org/10.5391/IJFIS.2022.22.3.223
Copyright © The Korean Institute of Intelligent Systems.
Dheo Prasetyo Nugroho, Sigit Widiyanto, and Dini Tri Wardani
Department of Information System Management, Gunadarma University, Depok, Indonesia
Correspondence to:Dheo Prasetyo Nugroho (dheoprasetyo.dp@gmail.com)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Examination of the technological development in agriculture reveals that not many applications use cameras to detect tomato ripeness; therefore, tomato maturity is still determined manually. Currently, technological advances and developments are occurring rapidly, and are, therefore, also inseparable from the agricultural sector. Object detection can help determining tomato ripeness. In this research, faster region-based convolutional neural network (Faster R-CNN), single shot multibox detector (SSD), and you only look once (YOLO) models were tested to recognize or detect tomato ripeness using input images. The model training process required 5 hours and produced a total loss value <0.5, and as the total loss became smaller, the predicted results improved. Tests were conducted on a training dataset, and average accuracy values of 99.55%, 89.3%, and 94.6% were achieved using the Faster R-CNN, SSD, and YOLO models, respectively.
Keywords: SSD, Faster R-CNN, YOLO, Object detection
Faster R-CNN architecture.
SSD architecture.
YOLO architecture.
Image classification, localization, and object detection.
Research methodology.
Image dataset.
Flowchart annotation process.
Annotation process.
Image annotation results using Faster R-CNN, SSD (left), and YOLO (right).
Flowchart of Faster RCNN training process.
Architecture of Faster R-CNN.
SSD architecture.
Flowchart of YOLO training process.
YOLO architecture.
Image prediction flowchart.
Training process on Google Colaboratory.
Total losses using (a) Faster R-CNN, (b) SSD, and (c) YOLO models.
Detection result.
Test result using webcam.
Table 1 . Image data collection.
Image | Number of images |
---|---|
Ripe tomatoes | 100 |
Half-ripe tomatoes | 100 |
Half-unripe tomatoes | 100 |
Unripe tomatoes | 100 |
Table 2 . mAP results.
Model | mAP |
---|---|
Faster R-CNN | 0.908 |
SSD | 0.858 |
YOLO | 0.778 |
Table 3 . Results of training.
Image | Faster R-CNN | SSD | YOLO |
---|---|---|---|
Half-ripe1 | 100 % | 78% | 100% |
Half-ripe2 | 100% | 97% | 100% |
Half-ripe3 | 100% | 100% | 99% |
Half-ripe4 | 99% | 65% | 100% |
Half-ripe5 | 100% | 99% | 99% |
Half-unripe1 | 98% | 62% | 63% |
Half-unripe2 | 100% | 97% | 93% |
Half-unripe3 | 100% | 80% | 53% |
Half-unripe4 | 98% | 98% | 99% |
Half-unripe5 | 100% | 98% | 95% |
Ripe1 | 100% | 100% | 100% |
Ripe2 | 100% | 97% | 99% |
Ripe3 | 100% | 98% | 97% |
Ripe4 | 100% | 98% | 100% |
Ripe5 | 99% | 51% | 98% |
Unripe1 | 99% | 83% | 100% |
Unripe2 | 100% | 100% | 98% |
Unripe3 | 99% | 100% | 100% |
Unripe4 | 99% | 87% | 100% |
Unripe5 | 100% | 98% | 99% |
Table 4 . Detection speed.
Resolution | Faster R-CNN | SSD | YOLO |
---|---|---|---|
1024 × 720 | 0.58 fps | 4.53 fps | 2.23 fps |
720 × 480 | 0.63 fps | 5.24 fps | 2.67 fps |
Laily Nur Qomariyati, Nurul Jannah, Suryo Adhi Wibowo, and Thomhert Suprapto Siadari
International Journal of Fuzzy Logic and Intelligent Systems 2024; 24(3): 194-202 https://doi.org/10.5391/IJFIS.2024.24.3.194Donho Nam and Seokwon Yeom
International Journal of Fuzzy Logic and Intelligent Systems 2020; 20(1): 43-51 https://doi.org/10.5391/IJFIS.2020.20.1.43Min-Hyuck Lee, and Seokwon Yeom
International Journal of Fuzzy Logic and Intelligent Systems 2018; 18(3): 182-189 https://doi.org/10.5391/IJFIS.2018.18.3.182Faster R-CNN architecture.
|@|~(^,^)~|@|SSD architecture.
|@|~(^,^)~|@|YOLO architecture.
|@|~(^,^)~|@|Image classification, localization, and object detection.
|@|~(^,^)~|@|Research methodology.
|@|~(^,^)~|@|Image dataset.
|@|~(^,^)~|@|Flowchart annotation process.
|@|~(^,^)~|@|Annotation process.
|@|~(^,^)~|@|Image annotation results using Faster R-CNN, SSD (left), and YOLO (right).
|@|~(^,^)~|@|Flowchart of Faster RCNN training process.
|@|~(^,^)~|@|Architecture of Faster R-CNN.
|@|~(^,^)~|@|SSD architecture.
|@|~(^,^)~|@|Flowchart of YOLO training process.
|@|~(^,^)~|@|YOLO architecture.
|@|~(^,^)~|@|Image prediction flowchart.
|@|~(^,^)~|@|Training process on Google Colaboratory.
|@|~(^,^)~|@|Total losses using (a) Faster R-CNN, (b) SSD, and (c) YOLO models.
|@|~(^,^)~|@|Detection result.
|@|~(^,^)~|@|Test result using webcam.