The advancement of precision agriculture heavily relies on the integration of AI and computer vision techniques to automate innovation farming tasks. The advent of UAVs, widely known as drones, has become an invaluable tool in various fields due to their ability to provide high-resolution, cost-effective and real-time data (
Goodchild, 2007). UAVs have brought about a paradigm shift in agricultural practices, particularly in the domain of orchard management and fruit tree cultivation
(Kumar et al., 2021). Their flexibility, rapid deployment and manoeuvrability make them invaluable in various applications, including environmental monitoring, precision agriculture, infrastructure inspection, mapping, security and surveillance
(Tuia et al., 2016). The significance of UAVs in image collection lies in their ability to provide high-resolution, real-time, cost-effective data and improved accessibility to challenging terrains, making them versatile tools across industries such as agriculture, disaster response and entertainment for a wide range of applications
(Kumar et al., 2021). The author
Girijalaxmi et al. (2024) developed an algorithm to calculate the distance between trees for agricultural applications. The integration of UAV technology in mango tree data collection stands as a testament to this transformation, ushering in an era of precision agriculture for one of the world’s most beloved fruits
(Ma et al., 2019).
The author
Houde et al. (2024) used the online mango tree videos to generate images for the development of a tree detection model. A robust UAVs based images dataset is crucial for the development of UAV and deep learning-based AI applications
(Ma et al., 2019). However, a major bottleneck in developing robust AI models for tree-level applications such as spraying, yield estimation and health monitoring is the lack of publicly available, high-resolution and well-annotated datasets, especially for fruit trees like mangoes, apple, citrus
etc. Existing agricultural datasets predominantly focus on common field crops and do not offer object-level annotations or imagery tailored to orchard environments. Even when such datasets exist, they are typically not captured using UAVs, making them unsuitable for aerial tree-level analysis. UAVs can generate large volumes of high-resolution images during a single flight. A robust database is essential to handle massive data efficiently and ensure quick and reliable retrieval
(Zhu et al., 2017). A well-designed database structure helps in organizing UAV-generated images systematically, making it easier to access specific datasets, locations, or time frames
(Sawat et al., 2016). This organization enhances data accessibility for analysis, research and decision-making, facilitating seamless integration with deep learning frameworks supporting tasks such as image annotation, training and validation
(Zhang et al., 2016). Additionally, it allows for the exploration of temporal trends, enhances data preprocessing for machine learning models and ensures the reliability and scalability necessary for the advancement of AI applications in fields like object detection, segmentation and environmental monitoring. UAVs often carry various sensors capturing different types of data, such as visual, infrared, or multispectral imagery
(Jose et al., 2021). Since UAVs can capture images at different time intervals, a database supporting temporal analysis allows researchers and analysts to track changes over time
(Jr et al., 2013). This is valuable for monitoring dynamic environments, such as urban development or natural resource changes.
To address this gap, we present a comprehensive UAV-based dataset of mango trees with YOLO-format annotations, specifically designed to support detection and segmentation tasks in real-world orchard settings.
UAVs, often known as drones, have significantly transformed various industries by providing advanced, high-resolution imaging capabilities and unmatched flexibility in operations. For our data collection, we used the UAV DJI Air 2S (Fig 1).
Literature review
The recent surge in studies exploring the use of unmanned aerial vehicles (UAVs) in various fields highlights their versatility and capability in data collection and analysis.
Khan et al. (2017) illustrate UAVs’ transformative role in urban planning and traffic management, providing a bird’s-eye view that surpasses traditional ground-based methods. UAVs equipped with high-resolution cameras can capture detailed imagery of urban traffic, processed using advanced computer vision algorithms for real-time monitoring and object detection.
UAVs were also used for remote sensing and various civil applications.
Shakhatreh et al. (2019) and
Mithra et al. (2021) provided a comprehensive overview of UAV uses, from image and video data collection to environmental monitoring and infrastructure assessment, addressing current challenges and future research directions. These studies reflect the expanding role of UAVs in various sectors, from urban development to agriculture and their intersection with advanced technologies like AI and machine learning. Integrating these tools and techniques paves the way for innovative solutions and enhanced efficiency in data collection, analysis and application across multiple domains.
Data collection overview
For the proposed work, the data was collected from the Bagalkot, Hubli, Belagavi and Dharwad districts of Karnataka state in the southern part of India and Dapoli, Ratnagiri district of Maharashtra state. The selection of these regions for UAV data collection is based on several factors, primarily the abundance of mango trees available in these regions
Vinita et al. (2022) and
Ausari et al., (2023). These regions experience a subtropical climate, which is conducive to mango cultivation. The warm temperatures and moderate rainfall during the monsoon season provide an ideal environment for mango trees to thrive and bear fruit
Ray et al. (2022) and
Gulati et al. (2021). These regions boast a rich agricultural landscape with extensive mango orchards and plantations. The fertile soil and availability of water resources support the cultivation of mango trees on a large scale.
Hardware and software used for data collection
Dronelink android-based mapping software is used over a Redme Note 11S cellphone for drone navigation, control and image capturing. DroneLink is a comprehensive software platform designed to streamline and enhance the process of aerial mapping and data collection using drones. It offers intuitive mission planning tools, real-time monitoring and advanced data processing capabilities. Fig 2 shows the map named “Dharwad2” captured at an altitude of 30 mts. The total covered area is 7.5 hectares, with a front overlap of 40% and a side overlap of 40%. In DroneLink, the “normal” pattern refers to a predefined flight path or trajectory designed to cover a specific area during aerial mapping missions. This pattern typically involves flying the drone in a systematic grid-like or zigzag pattern over the designated area, ensuring comprehensive coverage and consistent data collection. The parameters “front overlap 40% and “side overlap 40% specify the amount of overlap between consecutive images captured by the drone. Table 1 shows some common parameters used for all the images captured through the drone.
The dataset collected through the UAV comprises 3917 RGB images with dimensions of 5472 × 3648 pixels each, featuring a standard resolution of 72 dpi. The gimbal pitch was set to -90 degrees, capturing images with a downward perspective. The lens used had a focal length of 8.5mm, mounted on a sensor measuring 13.20 mm × 8.80 mm. The images were captured over a period spanning from January 14
th to March 24
th, 2024. These images occupy a disk space ranging from 10 MB to 14 MB per image, reflecting the high-resolution nature of the dataset. These images offer a comprehensive visual dataset for analysis and research. The consistent normal pattern of data collection and high-resolution imagery make this dataset valuable for deep learning-based AI applications and further study in fields like object detection, segmentation and landscape analysis.
Geographic locations
The UAV-collected dataset encompasses the rural and urban regions of Bagalkot, Hubli-Dharwad and Belagavi in Karnataka state and Dapoli, Ratnagiri in Maharashtra state of India. The geographical coordinates range from latitude X to Y and longitude A to B. Google Play Store GPS Map Camera, an Android-based open source application, is used to capture Latitude and Longitude values from the UAV Home-point locations which provides location information.
The maps depicting the data-collected area were crafted using QGIS, a powerful and versatile GIS software. In Fig 3, the green color dots from the Maharashtra and Karnataka districts illustrate the data-collected area, generated using QGIS and the spatial patterns of the observed phenomena. The color-coded layers in the map showcase variations in terrain characteristics, aiding in the interpretation of field observations. QGIS’s capabilities in geospatial analysis played a pivotal role in uncovering patterns and trends within the study area.
The maps depicting the data-collected area were crafted using QGIS, a powerful and versatile GIS software. QGIS allows for seamless integration of various data layers, including aerial imagery and geospatial features. The created maps provide a comprehensive visualization of the study area, highlighting specific points of interest and the spatial distribution of collected data.