Main Article Content
In 2018, the Indonesian fruit exports increased by 24% from the previous year. The surge in demand for tropical fruits from non-tropical countries is one of the contributing factors for this trend. Some of these countries have strict quality requirements – the poor level quality control of fruit is an obstacle in achieving greater export yield. This is because some exporters still use manual sorting processes performed by workers, hence the quality standard varies depending on the individual perception of the workers. Therefore, we need an intelligent system that is capable of automatic sorting according to the standard set. In this research, we propose a system that can classify fruit defects automatically. Faster R-CNN (FRCNN) architecture proposed as a solution to detect the level of defect on the surface of the fruit. There are three types of fruit that we research, its mangoes (sweet fragrant), lime, and pitaya fruit. Each fruit divided into three categories (i) Super, (ii) middle, (iii) and fruit defects. We exploit join detection and video tracking to calculate and determine the quality fruit in real-time. The datasets are taken in the field, then trained using the FRCNN Framework using the Tensorflow platform. We demonstrated that this system can classify fruit with an accuracy level of 88% (mango), 83% (lime), and 99% (pitaya), with an average computation cost of 0.0131 m/s. We can track and calculate fruit sequentially without using additional sensors and check the defect rate on fruit using the video streaming camera more accurately and with greater ease.
This work is licensed under a Creative Commons Attribution 4.0 International License.
Copyright Notice of Jurnal Ilmiah Kursor, Informatics Engineering, University of Trunojoyo Madura follows Creative Commons Attribution 4.0 International License
E. B. Esguerra and R. Rolle, Post-harvest management of mango for quality and safety assurance. United Nations: Food and Agriculture Organization(FAO), 2018.
S. CODEX, Standard for Pitahayas are classified in three classes defined. United Nations: Codex Committee on Fresh Fruits and Vegetables (CCFFV), 2011.
S. CODEX, Standard for Limes are classified in three classes defined. United Nations: Codex Committee on Fresh Fruits and Vegetables (CCFFV), 2011.
H. Basri, I. Syarif, and S. Sukaridhoto, “Faster R-CNN Implementation Method for Multi- Fruit Detection Using Tensorflow Platform,” 2018, no. October, pp. 1–4.
L. Ma, S. fadillah Umayah, S. Riyadi, C. Damarjati, and N. A. Utama, “Deep Learning Implementation using Convolutional Neural Network in Mangosteen Surface Defect Detection,” 2017, no. November, pp. 24–26.
K. N. Ranjit, H. K. Chethan, and C. Naveena, “Identification and Classification of Fruit Diseases,” vol. 6, no. 7, pp. 11–14, 2016.
I. Sa, Z. Ge, F. Dayoub, B. Upcroft, T. Perez, and C. McCool, “Deep fruits: A fruit detection system using deep neural networks,” Sensors (Switzerland), vol. 16, no. 8, 2016.
S. Jana and S. Basak, “Automatic Fruit Recognition from Natural Images using Color and Texture Features,” Devices Integr. Circuit, pp. 620–624, 2017.
M. Kaur and R. Sharma, “Quality Detection of Fruits by Using ANN Technique,” IOSR J. Electron. Commun. Eng. Ver. II, vol. 10, no. 4, pp. 2278–2834, 2015.
C. S. Nandi, B. Tudu, and C. Koley, “A machine vision-based maturity prediction system for sorting of harvested mangoes,” IEEE Trans. Instrum. Meas., vol. 63, no. 7, pp. 1722–1730, 2014.
S. Basak, “An Improved Bag-of-Features Approach for Object Recognition from Natural Images,” vol. 151, no. 10, pp. 5–11, 2016.
S. Nuske, S. Achar, T. Bates, S. Narasimhan, and S. Singh, “Yield Estimation in Vineyards by Visual Grape Detection (in process of publication),” 2011 IEEE Conf. Intell. Robot. Syst., pp. 2352–2358, 2011.
K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,” arXiv Technical Report, 2014. [Online]. Available: https://arxiv.org/abs/1409.1556. [Accessed: 14-Feb-2019].
L. Agilandeeswari, M. Prabukumar, and G. Shubham, “Automatic Grading System Mangoes Using Multiclass SVM Classifier,” Int. J. Pure Appl. Math., vol. 116, no. 23, pp. 515–523, 2017.
Indarto and Murinto, “Banana Fruit Detection Based on Banana Skin Image Features Using HSI Color Space Transformation Method,” JUITA, vol. 5, no. 1, pp. 15–21, 2017.
C. McCool, I. Sa, F. Dayoub, C. Lehnert, T. Perez, and B. Upcroft, “Visual Detection of Occluded Crop: For Automated Harvesting,” in Proc. IEEE Int. Conf. Robot. Autom, 2016, pp. 2506–2512.
T. Ojala, M. Pietikainen, and T. Maenpaa, “Multiresolution gray-scale and rotation invariant texture classiﬁcation with local binary patterns,” IEEE Trans. Pattern Anal. M ach. Intell, vol. 24, no. 7, pp. 971–987, 2002.
N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit, 2005, pp. 886–893.
S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” in Proc. Adv. Neural Inf. Process. Syst, 2015, pp. 91–99.
S. Chen, “Counting apples and oranges with deep learning: A data-driven approach,” IEEE Robot. Autom. Lett, vol. 2, no. 2, pp. 781–788, 2017.
Y. LeCun, B. Boser, J. S. Denker, and D. Henderson, Backpropagation Recognition, applied to handwritten zip code. 2015.