MultiCore+TPU Accelerated Multi-Modal TinyML for Livestock Behaviour Recognition

๐Ÿ“… 2025-04-10
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the low efficiency and deployment challenges of livestock behavior recognition and motion tracking in remote pasture environments with constrained network connectivity, this paper proposes a lightweight multimodal TinyML system that fuses accelerometer and edge vision data to enable real-time on-device behavior classification, image classification, and object detection. We introduce a novel MultiCore+TPU co-acceleration deployment paradigm, supporting collaborative inference across multiple devices under offline or low-bandwidth conditions, and jointly optimize the model, hardware, and communication stack. Through quantization-aware compression and TPU acceleration, the model size is reduced by 270ร—, end-to-end inference latency drops below 80 ms, and accuracy matches state-of-the-art methodsโ€”while maintaining stable operation on commercial microcontroller units (MCUs). This significantly enhances the practicality and scalability of intelligent livestock monitoring systems.

Technology Category

Application Category

๐Ÿ“ Abstract
The advancement of technology has revolutionised the agricultural industry, transitioning it from labour-intensive farming practices to automated, AI-powered management systems. In recent years, more intelligent livestock monitoring solutions have been proposed to enhance farming efficiency and productivity. This work presents a novel approach to animal activity recognition and movement tracking, leveraging tiny machine learning (TinyML) techniques, wireless communication framework, and microcontroller platforms to develop an efficient, cost-effective livestock sensing system. It collects and fuses accelerometer data and vision inputs to build a multi-modal network for three tasks: image classification, object detection, and behaviour recognition. The system is deployed and evaluated on commercial microcontrollers for real-time inference using embedded applications, demonstrating up to 270$ imes$ model size reduction, less than 80ms response latency, and on-par performance comparable to existing methods. The incorporation of the TinyML technique allows for seamless data transmission between devices, benefiting use cases in remote locations with poor Internet connectivity. This work delivers a robust, scalable IoT-edge livestock monitoring solution adaptable to diverse farming needs, offering flexibility for future extensions.
Problem

Research questions and friction points this paper is trying to address.

Develops TinyML-based livestock behavior recognition system
Fuses accelerometer and vision data for multi-modal tasks
Enables real-time edge inference with low latency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-modal TinyML for livestock behavior recognition
Accelerometer and vision data fusion
Real-time edge inference with low latency
๐Ÿ”Ž Similar Papers
No similar papers found.
Q
Qianxue Zhang
Medical AI Lab, Hebei Provincial Engineering Research Center for AI-Based Cancer Treatment Decision-Making, The First Hospital of Hebei Medical University, Shijiazhuang 050000, China; and the Computing Department, Imperial College London, London, UK.
Eiman Kanjo
Eiman Kanjo
Professor, Imperial College London
TinyMLEdge AIDecentralised AICollaborative & Distribuited AIPervasive Computing