Two-Stage Framework for Efficient UAV-Based Wildfire Video Analysis with Adaptive Compression and Fire Source Detection

📅 2025-08-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of real-time wildfire video analysis and ignition detection under stringent onboard computational constraints of unmanned aerial vehicles (UAVs), this paper proposes a lightweight two-stage collaborative framework. In the first stage, a policy-network-driven frame compression mechanism, augmented by future-frame-informed salient point prediction, efficiently eliminates redundant video segments. In the second stage, upon flame response triggering, a lightweight, modified YOLOv8 model performs precise fire source localization. This work introduces, for the first time, the integration of sequence-level salient point modeling with an adaptive compression-detection co-design architecture. Evaluations on FLAME, HMDB51, and Fire & Smoke datasets demonstrate that the framework maintains high classification accuracy (>92%), improves detection mAP by 3.7%, and reduces inference latency by 38% compared to baseline methods—achieving a significant balance between real-time performance and detection accuracy.

Technology Category

Application Category

📝 Abstract
Unmanned Aerial Vehicles (UAVs) have become increasingly important in disaster emergency response by enabling real-time aerial video analysis. Due to the limited computational resources available on UAVs, large models cannot be run independently for real-time analysis. To overcome this challenge, we propose a lightweight and efficient two-stage framework for real-time wildfire monitoring and fire source detection on UAV platforms. Specifically, in Stage 1, we utilize a policy network to identify and discard redundant video clips using frame compression techniques, thereby reducing computational costs. In addition, we introduce a station point mechanism that leverages future frame information within the sequential policy network to improve prediction accuracy. In Stage 2, once the frame is classified as "fire", we employ the improved YOLOv8 model to localize the fire source. We evaluate the Stage 1 method using the FLAME and HMDB51 datasets, and the Stage 2 method using the Fire & Smoke dataset. Experimental results show that our method significantly reduces computational costs while maintaining classification accuracy in Stage 1, and achieves higher detection accuracy with similar inference time in Stage 2 compared to baseline methods.
Problem

Research questions and friction points this paper is trying to address.

Efficient real-time wildfire video analysis on UAVs
Adaptive compression to reduce computational costs
Accurate fire source detection with lightweight models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Two-stage framework with adaptive video compression
Policy network with future frame information
Improved YOLOv8 model for fire localization
🔎 Similar Papers
No similar papers found.
Y
Yanbing Bai
Center for Applied Statistics, School of Statistics, Renmin University of China, Beijing 100872, China
Rui-Yang Ju
Rui-Yang Ju
National Taiwan University
Computer VisionHandwriting RecognitionDocument AnalysisMedical Image Processing
L
Lemeng Zhao
Center for Applied Statistics, School of Statistics, Renmin University of China, Beijing 100872, China
J
Junjie Hu
Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen 518129, China
J
Jianchao Bi
Gaoling School of Artificial Intelligence, Renmin University of China, Beijing 100872, China
Erick Mas
Erick Mas
Tohoku University
disaster sciencetsunami risk reductionagent-based modeling
Shunichi Koshimura
Shunichi Koshimura
International Research Institute of Disaster Science, Tohoku Univ.
Tsunami Science and EngineeringNumerical ModelingRemote SensingGeoinformatics