Research Article

Effects of Data Augmentation Methods on YOLO v5s: Application of Deep Learning with Pytorch for Individual Cattle Identification

Volume: 33 Number: 3 September 30, 2023
EN

Effects of Data Augmentation Methods on YOLO v5s: Application of Deep Learning with Pytorch for Individual Cattle Identification

Abstract

In this paper, we investigate the performance of the YOLO v5s (You Only Look Once) model for the identification of individual cattle in a cattle herd. The model is a popular method for real-time object detection, accuracy, and speed. However, since the videos obtained from the cattle herd consist of free space images, the number of frames in the data is unbalanced. This negatively affects the performance of the YOLOv5 model. First, we investigate the model performance on the unbalanced initial dataset obtained from raw images, then we stabilize the initial dataset using some data augmentation methods and obtain the model performance. Finally, we built the target detection model and achieved excellent model performance with an mAP (mean average precision) of 99.5% on the balanced dataset compared to the model on the unbalanced data (mAP of 95.8%). The experimental results show that YOLO v5s has a good potential for automatic cattle identification, but with the use of data augmentation methods, superior performance can be obtained from the model.

Keywords

References

  1. Altınbilek, H. F., & Kızıl, U. (2022). Identification of Paddy Rice Diseases Using Deep Convolutional Neural Networks. Yuzuncu Yıl University Journal of Agricultural Sciences, 32(4), 705-713. https://doi.org/10.29133/yyutbd.1140911
  2. Andrew W., Greatwood C., Burghardt T. (2017). Visual localisation and individual identification of holstein friesian cattle via deep learning. In Proceedings of the IEEE international conference on computer vision workshops. pp. 2850-2859.
  3. Andrew, W., Hannuna, S., Campbell, N., & Burghardt, T. (2016). Automatic individual holstein friesian cattle identification via selective local coat pattern matching in RGB-D imagery. Proceedings - International Conference on Image Processing, ICIP, 2016-August, 484–488. https://doi.org/10.1109/ICIP.2016.7532404
  4. Bati, C. T., & Ser, G. (2023). SHEEPFEARNET: Sheep fear test behaviors classification approach from video data based on optical flow and convolutional neural networks. Computers and Electronics in Agriculture, 204, 107540. https://doi.org/10.1016/J.COMPAG.2022.107540
  5. Bocaj, E., Uzunidis, D., Kasnesis, P., & Patrikakis, C. Z. (2020). On the Benefits of Deep Convolutional Neural Networks on Animal Activity Recognition. Proceedings of 2020 International Conference on Smart Systems and Technologies, SST 2020, 83–88. https://doi.org/10.1109/SST49455.2020.9263702
  6. Bochkovskiy, A., Wang, C.-Y., & Liao, H.-Y. M. (2020). YOLOv4: Optimal Speed and Accuracy of Object Detection. https://arxiv.org/abs/2004.10934v1
  7. Chen, Z., Wu, R., Lin, Y., Li, C., Chen, S., Yuan, Z., Chen, S., & Zou, X. (2022). Plant Disease Recognition Model Based on Improved YOLOv5. Agronomy 2022, Vol. 12, Page 365, 12(2), 365. https://doi.org/10.3390/AGRONOMY12020365
  8. Dac H.H., Gonzalez Viejo C., Lipovetzky N., Tongson E., Dunshea F. R., Fuentes S. (2022). Livestock Identification Using Deep Learning for Traceability. Sens. 22(21): 8256.

Details

Primary Language

English

Subjects

Agricultural, Veterinary and Food Sciences

Journal Section

Research Article

Early Pub Date

September 11, 2023

Publication Date

September 30, 2023

Submission Date

February 2, 2023

Acceptance Date

May 29, 2023

Published in Issue

Year 2023 Volume: 33 Number: 3

APA
Bati, C. T., & Ser, G. (2023). Effects of Data Augmentation Methods on YOLO v5s: Application of Deep Learning with Pytorch for Individual Cattle Identification. Yuzuncu Yıl University Journal of Agricultural Sciences, 33(3), 363-376. https://doi.org/10.29133/yyutbd.1246901

Cited By

Creative Commons License
Yuzuncu Yil University Journal of Agricultural Sciences by Van Yuzuncu Yil University Faculty of Agriculture is licensed under a Creative Commons Attribution 4.0 International License.