TY - CONF
T1 - Augmenting cattle tracking efficiency through monocular depth estimation
AU - Dickson, Lewis T.
AU - Davison, Christopher
AU - Michie, Craig
AU - McRobert, Ewan
AU - Atkinson, Robert
AU - Andonovic, Ivan
AU - Ferguson, Holly
AU - Dewhurst, Richard
AU - Briddock , Roger
AU - Brooking, Mark
AU - Pavlovic, Dejan
AU - Marko, Oskar
AU - Crnojevic, Vladimir
AU - Tachtatzis, Christos
PY - 2024/9/28
Y1 - 2024/9/28
N2 - We present a method for 3D cattle tracking and inter-camera pose transformation using depth information from monocular depth estimation with deep networks. Camera-based animal monitoring offers a minimally invasive and easily adaptable solution for tracking and welfare monitoring, relying solely on commercial RGB camera systems. However, environmental factors and inter-animal occlusion often hinder tracking efficacy and consistency. To address these challenges, we developed a pipeline to extract 3D point cloud data of individual cows in a straw-bedded calving yard environment, generating quasi-3D bounding boxes (x, y, z, height, width, θ), where θ is the polar angle. We then estimate the camera system extrinsic parameters by minimising the rotation, translation, and scale discrepancies between the apparent motion of animals across different frames of reference. This approach demonstrates a strong agreement between the 3D centroids of tracked animals in motion. Our work advances the development of algorithmic occlusion handling and object handover techniques in multi-camera systems, particularly pertinent to the high-occlusion, low-locomotion scenario of animals within barn environments.
AB - We present a method for 3D cattle tracking and inter-camera pose transformation using depth information from monocular depth estimation with deep networks. Camera-based animal monitoring offers a minimally invasive and easily adaptable solution for tracking and welfare monitoring, relying solely on commercial RGB camera systems. However, environmental factors and inter-animal occlusion often hinder tracking efficacy and consistency. To address these challenges, we developed a pipeline to extract 3D point cloud data of individual cows in a straw-bedded calving yard environment, generating quasi-3D bounding boxes (x, y, z, height, width, θ), where θ is the polar angle. We then estimate the camera system extrinsic parameters by minimising the rotation, translation, and scale discrepancies between the apparent motion of animals across different frames of reference. This approach demonstrates a strong agreement between the 3D centroids of tracked animals in motion. Our work advances the development of algorithmic occlusion handling and object handover techniques in multi-camera systems, particularly pertinent to the high-occlusion, low-locomotion scenario of animals within barn environments.
KW - precision farming
KW - monocular depth estimation
KW - tracking
KW - camera calibration
M3 - Paper
T2 - IEEE Conference on AgriFood Electronics
Y2 - 26 September 2024 through 28 September 2024
ER -