
pose model (image-based and keypoint-based). The
best pose estimation performance is obtained by the
YOLOv8-pose model, which demonstrates that com-
bining keypoint features and image features enhances
the pose estimation results. This work demonstrates
the capability of the proposed algorithm to accu-
rately recognize specific body parts using keypoint
detection, thereby providing a concurrent assessment
of pig-posture status. Overall, the results demon-
strate that combining image features and keypoint
features yields the most accurate pose estimation. The
YOLOv8-pose model consistently outperforms both
the MLP model and the ResNet-18 model, highlight-
ing the effectiveness of integrating multiple feature
types. The presented approach provides a promis-
ing foundation for future research aimed at detect-
ing more complex behaviors, such as social interac-
tions among pigs, further enhancing animal welfare
and monitoring capabilities.
ACKNOWLEDGEMENTS
This work is funded by the Dutch NWO project IM-
AGEN [P18-19 Project 1] of the Perspectief research
program. The German facility in Germany was of-
fered by Topigs Norsvin in Helvoirt, the Netherlands,
for conducting the video recordings. We are grate-
ful to the researchers and student assistants from Wa-
geningen University & Research and Eindhoven Uni-
versity of Technology for their help with tracking
and behavior ground-truth annotations. Additionally,
we acknowledge the Genes2Behave project [321409
- IPNÆRINGSLIV20, G2B] for its contribution in
sharing the Norwegian dataset.
REFERENCES
Cao, Z., Hidalgo Martinez, G., Simon, T., Wei, S., and
Sheikh, Y. A. (2019). Openpose: Realtime multi-
person 2d pose estimation using part affinity fields.
IEEE Transactions on Pattern Analysis and Machine
Intelligence.
Fang, H.-S., Li, J., Tang, H., Xu, C., Zhu, H., Xiu, Y., Li,
Y.-L., and Lu, C. (2022). Alphapose: Whole-body
regional multi-person pose estimation and tracking in
real-time. IEEE Transactions on Pattern Analysis and
Machine Intelligence.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 770–778.
Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M.,
and Schiele, B. (2016). Deepercut: A deeper, stronger,
and faster multi-person pose estimation model. In
Computer Vision–ECCV 2016: 14th European Con-
ference, Amsterdam, The Netherlands, October 11-14,
2016, Proceedings, Part VI 14, pages 34–50. Springer.
Jocher, G., Chaurasia, A., and Qiu, J. (2023). Ultralytics
YOLO Docs.
Lee, J., Jin, L., Park, D., and Chung, Y. (2016). Auto-
matic recognition of aggressive behavior in pigs using
a kinect depth sensor. Sensors, 16(5):631.
Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P.,
Ramanan, D., Doll
´
ar, P., and Zitnick, C. L. (2014).
Microsoft coco: Common objects in context. In Euro-
pean conference on computer vision, pages 740–755.
Springer.
Niu, K. and Yan, Y. (2023). A small-object-detection model
based on improved yolov8 for uav aerial images. In
2023 2nd International Conference on Artificial In-
telligence and Intelligent Information Processing (AI-
IIP), pages 57–60. IEEE.
Racewicz, P., Ludwiczak, A., Skrzypczak, E.,
Składanowska-Baryza, J., Biesiada, H., Nowak,
T., Nowaczewski, S., Zaborowicz, M., Stanisz,
M., and
´
Sl
´
osarz, P. (2021). Welfare health and
productivity in commercial pig herds. Animals,
11(4):1176.
Rosenblatt, F. (1958). The perceptron: a probabilistic model
for information storage and organization in the brain.
Psychological review, 65(6):386.
Toshev, A. and Szegedy, C. (2014). Deeppose: Human pose
estimation via deep neural networks. In Proceedings
of the IEEE conference on computer vision and pat-
tern recognition, pages 1653–1660.
Volkmann, N., Zelenka, C., Devaraju, A. M., Br
¨
unger, J.,
Stracke, J., Spindler, B., Kemper, N., and Koch, R.
(2022). Keypoint detection for injury identification
during turkey husbandry using neural networks. Sen-
sors, 22(14).
Wang, C.-Y., Liao, H.-Y. M., Wu, Y.-H., Chen, P.-Y., Hsieh,
J.-W., and Yeh, I.-H. (2020). Cspnet: A new backbone
that can enhance learning capability of cnn. In Pro-
ceedings of the IEEE/CVF conference on computer
vision and pattern recognition workshops, pages 390–
391.
Wutke, M., Heinrich, F., Das, P. P., Lange, A., Gentz,
M., Traulsen, I., Warns, F. K., Schmitt, A. O., and
G
¨
ultas, M. (2021). Detecting animal contacts—a deep
learning-based pig detection and tracking approach
for the quantification of social contacts. Sensors,
21(22):7512.
Yang, Q. and Xiao, D. (2020). A review of video-based
pig behavior recognition. Applied Animal Behaviour
Science, 233:105146.
Zhuang, X., Bi, M., Guo, J., Wu, S., and Zhang, T. (2018).
Development of an early warning algorithm to detect
sick broilers. Computers and Electronics in Agricul-
ture, 144:102–113.
VISAPP 2025 - 20th International Conference on Computer Vision Theory and Applications
474