toplogo
Sign In

Automated Detection and Sizing of Immature Green Apples in Commercial Orchards using YOLOv8 and 3D Shape Fitting Techniques


Core Concepts
Accurate detection and sizing of immature green apples during early growth stages is crucial for yield prediction, pest management, and crop-load management in commercial orchards. This study demonstrates the feasibility of using the state-of-the-art YOLOv8 object detection and instance segmentation algorithm combined with 3D shape fitting techniques to efficiently determine the size of immature green apples in a real-world orchard environment.
Abstract
This study aimed to develop an automated detection and sizing method for immature green apples (fruitlets) in a commercial orchard environment using the YOLOv8 object detection model and 3D shape fitting techniques. The key highlights and insights are: The YOLOv8m-seg model achieved the highest AP@0.5 and AP@0.75 scores of 0.94 and 0.91, respectively, for immature green apple detection and segmentation. The ellipsoid fitting technique on 3D point clouds from the Microsoft Azure Kinect camera achieved an RMSE of 2.35 mm, MAE of 1.66 mm, MAPE of 6.15 mm, and an R-squared value of 0.9 in estimating the size of apple fruitlets. Partial occlusion caused by leaves and branches, as well as low light or shadow conditions, resulted in some errors in accurately delineating and sizing green apples using the YOLOv8-based segmentation technique, particularly in fruit clusters. The size estimation technique performed better on the images acquired with the Microsoft Azure Kinect camera compared to the Intel RealSense D435i camera, as evident from the lower RMSE, MAE, and higher R-squared values. This study demonstrated the feasibility of accurately sizing immature green fruit in early growth stages using the combined 3D sensing and shape-fitting technique, which shows promise for improved precision agricultural operations such as optimal crop-load management in orchards.
Stats
The YOLOv8m-seg model achieved an AP@0.5 of 0.94 and an AP@0.75 of 0.91 for immature green apple detection and segmentation. Using the ellipsoid fitting technique on images from the Microsoft Azure Kinect, the study achieved an RMSE of 2.35 mm, MAE of 1.66 mm, MAPE of 6.15 mm, and an R-squared value of 0.9 in estimating the size of apple fruitlets. For the Intel RealSense D435i camera, the ellipsoid fitting technique achieved an RMSE of 9.65 mm, MAE of 7.8 mm, MAPE of 29.48 mm, and an R-squared value of 0.77 in estimating the size of apple fruitlets.
Quotes
"Accurate information on number and size of fruit during this stage enables farmers to strategize and prepare for harvest and post-harvest logistics including the workforce, equipment, and storage requirements." "Accurate green fruit size estimates can help farmers predict future market reception of their crop. For instance, optimal size apples may go for higher prices than small or large apples."

Deeper Inquiries

How can the YOLOv8 model be further improved to address the challenges of partial occlusion and variable lighting conditions in the orchard environment?

To enhance the YOLOv8 model's performance in addressing challenges like partial occlusion and variable lighting conditions in orchard environments, several strategies can be implemented: Data Augmentation: Increasing the diversity of the training dataset by including images with varying levels of occlusion and lighting conditions can help the model learn to adapt to these challenges. Augmenting the dataset with artificially occluded images and images captured under different lighting scenarios can improve the model's robustness. Advanced Preprocessing Techniques: Implementing advanced preprocessing techniques such as contrast enhancement, shadow removal, and noise reduction can help improve the quality of input images, making it easier for the model to detect and segment immature green apples accurately. Transfer Learning: Leveraging transfer learning by fine-tuning the pre-trained YOLOv8 model on a dataset specifically focused on partial occlusion and variable lighting conditions can help the model learn to handle these challenges more effectively. Post-Processing Techniques: Implementing post-processing techniques like morphological operations to refine the segmentation results can help in dealing with partial occlusion and improving the accuracy of fruit detection. Ensemble Methods: Combining the outputs of multiple YOLOv8 models trained with different strategies or on different subsets of data can help in improving the model's performance under challenging conditions.

How can the insights from this study on automated green fruit detection and sizing be extended to other tree fruit crops to support precision agriculture and crop management practices?

The insights gained from this study on automated green fruit detection and sizing can be extended to other tree fruit crops by: Dataset Expansion: Creating a diverse dataset that includes images of different tree fruit crops at various growth stages can help in training models that are specific to each crop. Model Generalization: Ensuring that the models developed for green fruit detection and sizing are generalizable to other tree fruit crops by considering factors like fruit shape, size, and color specific to each crop. Sensor Compatibility: Testing the developed models with different RGB-D sensors to ensure compatibility with the sensors commonly used in orchard environments for various tree fruit crops. Adaptation to Crop-Specific Challenges: Considering the unique challenges faced by different tree fruit crops, such as varying canopy structures, fruit shapes, and growth patterns, and adapting the detection and sizing models accordingly. Integration with Precision Agriculture Technologies: Integrating the automated detection and sizing systems with other precision agriculture technologies like robotic platforms, drones, and IoT devices to enable real-time monitoring and decision-making for crop management practices.

What other 3D sensing technologies or data fusion techniques could be explored to enhance the accuracy and robustness of immature fruit size estimation?

To enhance the accuracy and robustness of immature fruit size estimation, the following 3D sensing technologies and data fusion techniques could be explored: LiDAR (Light Detection and Ranging): LiDAR technology can provide high-resolution 3D point cloud data, enabling precise measurements of fruit size and shape in orchard environments. Structured Light Scanning: Utilizing structured light scanning techniques can enhance the accuracy of 3D reconstructions of fruit shapes, especially for irregularly shaped fruits. Time-of-Flight (ToF) Cameras: Time-of-Flight cameras can capture depth information quickly and accurately, aiding in the estimation of fruit size based on 3D data. Multi-Sensor Fusion: Integrating data from multiple sensors such as RGB cameras, LiDAR, and ToF cameras through sensor fusion techniques can provide a more comprehensive and accurate representation of fruit size and shape. Machine Learning-Based Fusion: Implementing machine learning algorithms for data fusion can help in combining information from different sensors to improve the accuracy of fruit size estimation. By exploring these 3D sensing technologies and data fusion techniques, the accuracy and robustness of immature fruit size estimation can be significantly enhanced, leading to more efficient precision agriculture practices.
0
visual_icon
generate_icon
translate_icon
scholar_search_icon
star