- Data Descriptor
- Open access
- Published: 29 January 2026
Scientific Data , Article number: (2026) Cite this article
We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.
Abstract
Efficient hay bale detection and counting are essential tasks within modern precision agriculture, significantly impacting yield estimation, logistics, and sustainable resource management. To address current limitations in dataset quality and en…
- Data Descriptor
- Open access
- Published: 29 January 2026
Scientific Data , Article number: (2026) Cite this article
We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.
Abstract
Efficient hay bale detection and counting are essential tasks within modern precision agriculture, significantly impacting yield estimation, logistics, and sustainable resource management. To address current limitations in dataset quality and environmental representation, we introduce BaleUAVision, a comprehensive dataset consisting of 2,599 high-resolution RGB images, each containing numerous human-annotated hay bales. Captured by Unmanned Aerial Vehicles (UAVs) across 16 diverse agricultural fields in Northern Greece, the dataset includes varying flight altitudes (50–100 meters), diverse speeds (3.7–5 m/s), and overlapping strategies to ensure robust data representation. BaleUAVision provides rich annotations through polygon-based semantic segmentation in multiple formats (COCO, CSV, JSON, YOLO, segmentation masks) and high-quality orthomosaics for precise spatial analysis. Technical validation demonstrated the dataset’s effectiveness in training robust hay bale detection models using YOLOv11, achieving high precision and recall under varying geographic and altitude conditions. Specifically, the dataset supported effective generalization across geographically distinct areas (Xanthi and Drama regions) and varying altitudes, highlighting its utility in real-world UAV operations. The dataset and supplementary tools, scripts, and analyses are publicly available on Zenodo and GitHub respectively, following FAIR principles to support wide-reaching applicability within the research community.
Data availability
All data comprising the BaleUAVision dataset are publicly available in the Zenodo repository29.
Code availability
All related code, including data statistics, insights, usage python scripts, and indicative examples are provided at the living page of the project30.
References
Henning, J. & Lawrence, L. Production and Management of Hay and Haylage (Academic Press, 2025). 1.
Pölönen, I., Suokannas, A. & Juntunen, A. Cloud-based approach for tracking and monitoring of hay bales in smart agriculture. Technology Innovation Management Review 11 (2021). 1.
Wolfert, S., Ge, L., Verdouw, C. & Bogaardt, M.-J. Big data in smart farming–a review. Agricultural Systems 153, 69–80 (2017).
Mulla, D. J. Twenty five years of remote sensing in precision agriculture: Key advances and remaining knowledge gaps. Biosystems Engineering 114, 358–371 (2013).
Zamar, D. S., Gopaluni, B. & Sokhansanj, S. A constrained k-means and nearest neighbor approach for route optimization in the bale collection problem. In IFAC-PapersOnLine, vol. 50, 12125–12130 (2017). 1.
Tsouros, D. C., Bibi, S. & Sarigiannidis, P. G. A review on uav-based applications for precision agriculture. Information 10, 349 (2019).
Del Cerro, J. et al. Unmanned aerial vehicles in agriculture: A survey. Agronomy 11, 203 (2021).
Raptis, E. K. et al. Cofly: An automated, ai-based open-source platform for uav precision agriculture applications. SoftwareX 23, 101414 (2023).
Krestenitis, M. et al. Overcome the fear of missing out: Active sensing uav scanning for precision agriculture. Robotics and Autonomous Systems 172, 104581 (2024).
Weyler, J. et al. Joint plant and leaf instance segmentation on field-scale uav imagery. IEEE Robotics and Automation Letters 7, 3787–3794 (2022).
Chebrolu, N. et al. Agricultural robot dataset for plant classification, localization and mapping on sugar beet fields https://www.ipb.uni-bonn.de/data/sugarbeets2016/ (2017). 1.
Krestenitis, M. et al. Cofly-weeddb: A uav image dataset for weed detection and species identification https://zenodo.org/records/6697343 (2022). 1.
Mujkic, E., Christiansen, M. P. & Ravn, O. Object detection for agricultural vehicles: Ensemble method based on hierarchy of classes. Sensors 23, 7285 (2023).
Kamilaris, A. & Prenafeta-Boldú, F. X. Deep learning in agriculture: A survey. Computers and Electronics in Agriculture 147, 70–90 (2018).
Yamada, W., Zhao, W. & Digman, M. Automated bale mapping using machine learning and photogrammetry. Remote Sensing 13, 4675 (2021).
Zhao, W. et al. Augmenting crop detection for precision agriculture with deep visual transfer learning-a case study of bale detection. Remote Sensing 13, 23 (2020).
Xu, J. et al. Harvestnet: A dataset for detecting smallholder farming activity using harvest piles and remote sensing https://figshare.com/s/45a7b45556b90a9a11d2 (2024). 1.
Kirillov, A. et al. Segment anything. In Proceedings of the IEEE/CVF international conference on computer vision, 4015–4026 (2023). 1.
Osco, L. P. et al. The segment anything model (sam) for remote sensing applications: From zero to one shot. International Journal of Applied Earth Observation and Geoinformation 124, 103540 (2023).
Gui, B., Bhardwaj, A. & Sam, L. Evaluating the efficacy of segment anything model for delineating agriculture and urban green spaces in multiresolution aerial and spaceborne remote sensing images. Remote sensing 16, 414 (2024).
Li, Y., Wang, D., Yuan, C., Li, H. & Hu, J. Enhancing agricultural image segmentation with an agricultural segment anything model adapter. Sensors 23, 7884 (2023).
Wasil, M., Drak, A., Penfold, B., Scarton, L., Johenneken, M., Asteroth, A., Houben, S. Parameter-efficient fine-tuning of vision foundation model for forest floor segmentation from uav imagery. arXiv preprint arXiv:2505.08932 (2025). 1.
Tripathy, P., Baylis, K., Wu, K., Watson, J. & Jiang, R. Investigating the segment anything foundation model for mapping smallholder agriculture field boundaries without training labels. arXiv preprint arXiv:2407.01846 (2024). 1.
Ye, Z. et al. A comparison between pixel-based deep learning and object-based image analysis (obia) for individual detection of cabbage plants based on uav visible-light images. Computers and Electronics in Agriculture 209, 107822 (2023).
Horning, N. et al. Mapping of land cover with open-source software and ultra-high-resolution imagery acquired with unmanned aerial vehicles. Remote Sensing in Ecology and Conservation 6, 487–497 (2020).
Lottes, P. et al. Uav-based crop and weed classification for smart farming. In 2017 IEEE International Conference on Robotics and Automation (ICRA) (2017). 1.
Karatzinis, G. D. et al. Towards an integrated low-cost agricultural monitoring system with unmanned aircraft system. In 2020 International Conference on Unmanned Aircraft Systems (ICUAS) (2020). 1.
Wilkinson, M. D. et al. The fair guiding principles for scientific data management and stewardship. Scientific Data 3, 1–9 (2016).
Karatzinis, G., Gkelios, S. & Kapoutsis, A. Baleuavision: Hay bales uav captured dataset https://doi.org/10.5281/zenodo.15304715 (2025). 1.
Karatzinis, G., Gkelios, S. & Kapoutsis, A. Baleuavision: High-resolution uav dataset for automated hay bale detection and counting. GitHub repository Available at https://github.com/georkara/BaleUAVision (2025). 1.
Ravi, N. et al. Sam 2: Segment anything in images and videos. arXiv preprint arXiv:2408.00714 (2024). 1.
Minderer, M., Gritsenko, A. & Houlsby, N. Scaling open-vocabulary object detection. Advances in Neural Information Processing Systems 36, 72983–73007 (2023).
Acknowledgements
This research was funded by European Union’s Horizon Europe Innovation Action iDriving, under Grant Agreement No 101147004. Views and opinions expressed are, however, those of the authors only and do not necessarily reflect those of the European Union or CINEA. Neither the European Union nor the granting authority can be held responsible for them.
Author information
Authors and Affiliations
Information Technologies Institute (ITI), Centre of Research and Technology Hellas (CERTH), Thessaloniki, Greece
Georgios D. Karatzinis, Socratis Gkelios & Athanasios Ch. Kapoutsis
Authors
- Georgios D. Karatzinis
- Socratis Gkelios
- Athanasios Ch. Kapoutsis
Contributions
Conceptualization, G.K.; data acquisition, G.K.; data annotation G.K.; methodology, G.K. and S.G.; technical validation, G.K. and S.G.; formal analysis, G.K. and S.G.; writing—original draft preparation, G.K., S.G. and A.K.; writing—review and editing, G.K., S.G. and A.K.; visualization, G.K., S.G. and A.K. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Correspondence to Georgios D. Karatzinis.
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Karatzinis, G.D., Gkelios, S. & Kapoutsis, A.C. BaleUAVision: Hay Bales UAV Captured Dataset. Sci Data (2026). https://doi.org/10.1038/s41597-026-06622-8
Received: 09 May 2025
Accepted: 13 January 2026
Published: 29 January 2026
DOI: https://doi.org/10.1038/s41597-026-06622-8