
6 CONCLUSIONS
In this paper, we proposed a method for the multi-
camera tracking of dairy cows that utilizes the homog-
raphy error for selecting reliable bounding boxes. Ex-
periments on actual scenes showed that the proposed
method achieved high accuracy for two different barn
environments; even when the camera shooting an-
gles were unaligned, the tracking accuracy was im-
proved by appropriately selecting reliable bounding
boxes using the homography error. In addition, ex-
periments confirmed that evaluating homography er-
ror is effective for appearance-feature based tracking
methods, not just location-based methods.
Future work includes verifying our method can
handle different camera configurations and arrange-
ment.
ACKNOWLEDGEMENTS
The authors thank the members of Obihiro University
of Agriculture and Veterinary Medicine and Tsuchiya
Manufacturing Co. Ltd for helpful discussions and
for providing the video data of barn.
REFERENCES
Akizawa, K., Yamamoto, Y., and Taniguchi, Y. (2022).
Dairy cow tracking across multiple cameras with
shared field of views. Technical Report 299, The in-
stitute of image electronics engineers of Japan. (in
Japanese).
Aou, S., Yamamoto, Y., Nakamura, K., and Taniguchi, Y.
(2024). Multi-camera tracking of dairy cows using
rotated bounding boxes. In Proc. IWAIT.
Bernardin, K. and Stiefelhagen, R. (2008). Evaluating mul-
tiple object tracking performance: the clear mot met-
rics. EURASIP journal on image and video process-
ing, 2008:1–10.
Bewley, A., Ge, Z., Ott, L., Ramos, F., , and Upcroft, B.
(2016). Simple online and realtime tracking. In Proc.
ICIP.
Bredereck, M., Jiang, X., K
¨
orner, M., and Denzler, J.
(2012). Data association for multi-object tracking-by-
detection in multi-camera networks. In 2012 Sixth In-
ternational Conference on Distributed Smart Cameras
(ICDSC), pages 1–6.
Derek, G., Zach
´
ar, G., Wang, Y., Ji, J., Nice, M., Bunting,
M., and Barbour, W. W. (2024). So you think you can
track? In Proc. WACV.
Eshel, R. and Moses, Y. (2008). Homography based multi-
ple camera detection and tracking of people in a dense
crowd. In 2008 IEEE Conference on Computer Vision
and Pattern Recognition, pages 1–8.
Fleuret, F., Berclaz, J., Lengagne, R., and Fua, P. (2008).
Multicamera people tracking with a probabilistic oc-
cupancy map. IEEE Transactions on Pattern Analysis
and Machine Intelligence, 30(2):267–282.
Hu, K. (2020). YOLOV5 for oriented object detection.
https://github.com/hukaixuan19970627/yolov5
obb.
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Im-
agenet classification with deep convolutional neural
networks. Advances in neural information processing
systems, 25.
Mar, C. C., Zin, T. T., Tin, P., Honkawa, K., Kobayashi,
I., and Horii, Y. (2023). Cow detection and tracking
system utilizing multi-feature tracking algorithm. Sci-
entific reports, 13(17423).
Qian, Q., Shang, L., Sun, B., Hu, J., Li, H., and Jin, R.
(2019). Softtriple loss: Deep metric learning with-
out triplet sampling. In Proceedings of the IEEE/CVF
international conference on computer vision, pages
6450–6458.
Ristani, E., Solera, F., Zou, R., Cucchiara, R., and Tomasi,
C. (2016). Performance measures and a data set for
multi-target, multi-camera tracking. In Proc. ECCV,
pages 17–35. Springer.
Ristani, E. and Tomasi, C. (2018). Features for multi-target
multi-camera tracking and re-identification. In Proc.
CVPR, pages 6036–6046.
Shirke, A., Saifuddin, A., Luthra, A., Li, J., Williams, T.,
Hu, X., Kotnana, A., Kocabalkanli, O., Ahuja, N.,
Green-Miller, A., Condotta, I., Dilger, R. N., and
Caesar, M. (2021). Tracking grow-finish pigs across
large pens using multiple cameras. arXiv preprint
arXiv:2111.10971.
Ultralytics (2024). Oriented Bounding Boxes Object Detec-
tion. https://docs.ultralytics.com/ja/tasks/obb/.
Yamamoto, Y., Akizawa, K., Aou, S., and Taniguchi, Y.
(2024). Entire-barn dairy cow tracking framework for
multi-camera systems. Computers and Electronics in
Agriculture, 229.
Zhang, Z. (2000). A flexible new technique for camera cal-
ibration. IEEE transactions on pattern analysis and
machine intelligence, 22(11):1330–1334.
VISAPP 2025 - 20th International Conference on Computer Vision Theory and Applications
682