In recent years, the frequent occurrence of high-rise building fires has posed significant challenges due to rapid fire spread, multiple propagation paths, and the tendency to form three-dimensional fires, leading to substantial economic losses and casualties. These challenges demand advanced capabilities from firefighters and existing firefighting equipment. Traditional firefighting equipment often suffers from limitations such as inability to access, deploy, or reach sufficient heights, exacerbating the difficulty of rescue operations in high-rise structures. Over the past decade, statistics indicate that high-rise fires have resulted in thousands of incidents, hundreds of casualties, and billions in direct economic losses. Conventional aerial ladder trucks, typically operating at heights of 20–50 meters, are inadequate for buildings exceeding 80–100 meters, where deployment becomes impractical. To address these deficiencies, the development of rapid and precise firefighting solutions for high-rise buildings has become an urgent priority. Unmanned aerial vehicles (UAVs), or fire UAVs, offer a promising alternative due to their unrestricted altitude capabilities, quick response times, and adaptability to challenging environments. In this context, I focus on the visual algorithm research for fire UAVs, specifically targeting flame detection and tracking in high-rise fire scenarios, which is critical for effective灭火 operations.
The integration of fire UAVs into firefighting leverages their cost-effectiveness, rapid deployment, operational simplicity, and high accessibility. However, traditional flight controllers in fire UAVs often face issues such as GPS interference from high temperatures, leading to failures in remote target aiming. To overcome this, I propose a combined algorithm based on Kalman filtering and mean shift for fast and accurate flame target recognition. This algorithm utilizes a Kalman filter for auxiliary position prediction, mitigating problems like target loss or misjudgment caused by significant changes in target size or occlusion. Experimental results demonstrate that this approach reduces visual recognition failure rates in high-rise fire UAV applications, with stability, precision, and anti-interference capabilities meeting operational requirements. This advancement holds particular significance for extinguishing initial-stage fires, potentially reducing casualties and economic losses. Below, I delve into the details of this research, covering background, algorithm principles, system design, experiments, and conclusions, all from my perspective as a researcher in this field.

The application of fire UAVs in firefighting has evolved significantly over the years. Internationally, early experiments date back to 1979 with UAVs used for aerial photogrammetry. By the 1980s and 1990s, U.S. research institutions made strides in UAV remote sensing image processing, applying成果 to agriculture monitoring, military reconnaissance, and real-time forest fire detection. Projects like Stanford University’s Hummingbird utilized GPS-automated helicopters for visual target localization and tracking. Today, small, low-cost UAV-camera systems serve as platforms for testing autonomous ground target recognition, obstacle avoidance, and multi-UAV协同跟踪. Advanced fire UAV systems, such as the MQ-1 Predator and Eagle Eye, provide long-duration, real-time monitoring and tracking capabilities. Domestically, progress includes 3D reconstruction of disaster scenes using low-altitude imagery and UAV flight data, as well as dynamic target detection基于深度卷积神经网络 deployed on embedded systems like NVIDIA Jetson TX1. The trend is toward multi-UAV cooperation and real-time tracking algorithms, with studies achieving autonomous control and collaborative decision-making in fire UAV fleets. These developments underscore the potential of fire UAVs in enhancing firefighting efficacy, especially for high-rise buildings where traditional methods fall short.
In target recognition algorithms for fire UAVs, the mean shift tracking algorithm is computationally simple and offers good real-time performance in简单场景 with minimal scale changes or background interference. However, it tends to lose track when new灰度 regions appear continuously in the target area, a common issue in high-rise fire scenarios where flame targets exhibit large-scale variations. To address this, I combine mean shift with Kalman filtering. The Kalman filter aids in position prediction, improving robustness against target loss due to size changes or prolonged occlusion. By estimating运动状态, the Kalman filter helps maintain tracking accuracy even when targets are temporarily obscured. This combined approach is particularly suited for fire UAV operations, where environmental factors like heat and smoke can disrupt visual inputs. The Kalman filter model is based on Gaussian distributions for state and observation noise, consisting of prediction and update stages. The state equation is given by:
$$x_k = A_k x_{k-1} + w_k$$
where \(x_k\) is the system state at time \(k\), \(A_k\) is the state transition matrix, and \(w_k\) is system noise. The observation equation is:
$$y_k = B_k x_k + v_k$$
where \(y_k\) is the measurement at time \(k\), \(B_k\) is the observation matrix, and \(v_k\) is observation noise. Both noises are assumed to be Gaussian white noise. The prediction stage estimates the current state based on the previous state:
$$\hat{x}_{k|k-1} = A_k \hat{x}_{k-1|k-1}$$
with covariance:
$$P_{k|k-1} = A_k P_{k-1|k-1} A_k^T + Q_{k-1}$$
where \(Q_{k-1}\) is the process noise covariance. The update stage incorporates measurements to refine the estimate:
$$\hat{x}_{k|k} = \hat{x}_{k|k-1} + K_k (y_k – B_k \hat{x}_{k|k-1})$$
with Kalman gain \(K_k\) calculated as:
$$K_k = \frac{P_{k|k-1} B_k^T}{B_k P_{k|k-1} B_k^T + R_k}$$
where \(R_k\) is the measurement noise covariance. The updated covariance is:
$$P_{k|k} = (I – K_k B_k) P_{k|k-1}$$
This model enables precise state estimation, which I integrate with mean shift for enhanced tracking in fire UAV applications. The mean shift algorithm iteratively shifts a kernel to the mode of a distribution, defined for target tracking by comparing histograms. Given a target histogram \(q\) and candidate histogram \(p(y)\), the similarity is measured using the Bhattacharyya coefficient. The combined algorithm步骤 are as follows: First, read the initial video frame and perform image preprocessing. Second, extract前景 from consecutive frames and judge火焰 targets based on color features. Third, use the Kalman filter to predict the initial target position \(y_0\) in the current frame and compute the target histogram around \(y_0\). Fourth, detect external contours near \(y_0\), calculate the centroid \(y_1\), and apply mean shift iteration to搜索 the target position \(y_1\). Compute the candidate histogram around \(y_1\) and evaluate similarity; if \(\|y_1 – y_0\| < \epsilon\), stop iteration; otherwise, update \(y_0\) and continue. Fifth, repeat for subsequent frames until completion. This integration allows the fire UAV to maintain tracking despite动态变化 in flame targets.
The system framework for the high-rise fire UAV is designed to handle challenging environments. It comprises several key components, as summarized in Table 1, which outlines the无人机平台 and its subsystems. The fire UAV platform features a protective shell with high-temperature insulation coatings to withstand heat radiation. It is equipped with a millimeter-wave collision avoidance radar for precise distance sensing in high-temperature气流, ensuring safe navigation near buildings. The control and information processing system includes a flight controller, artificial intelligence system, and machine vision module. Traditional flight controllers struggle with GPS interference and communication link failures, but I employ an NVIDIA TEGRA TX1 AI platform to run the improved Kalman滤波算法 for accurate flame targeting and灭火弹 deployment. The system框架 is illustrated in the inserted image, showing the integration of hardware and software elements. Data transmission systems, network modules,灭火弹发射系统, and battery management are all coordinated to support autonomous operations. This design emphasizes reliability and precision, critical for fire UAV missions in high-rise settings where human intervention is limited.
| Component | Description | Function in Fire UAV |
|---|---|---|
| UAV Platform | High-temperature resistant shell with insulation coating | Protects against heat radiation and ensures durability |
| Flight Controller | Integrated with AI and machine vision模组 | Controls UAV flight and processes visual data |
| Machine Vision Module | Cameras and sensors for image capture | Detects and tracks火焰 targets using algorithms |
| AI System | NVIDIA TEGRA TX1 platform | Runs Kalman滤波 and mean shift algorithms for targeting |
| Millimeter-wave Radar | 防撞 radar for distance sensing | Prevents collisions and measures building proximity |
| Fire Extinguishing System | 灭火弹发射机制 | Deploys extinguishing agents accurately at targets |
| Data Transmission | Network modules for communication | Links UAV to ground station for real-time control |
| Battery Management | Power system with efficient energy use | Ensures extended operational time for fire UAV |
To validate the combined algorithm, I conducted experiments using a Tiantu M6消防无人机 equipped with the NVIDIA TEGRA TX1 AI platform. The aiming system operated at 5.8 GHz with an error within \(20 \pm 0.5\) meters, suitable for high-rise applications. The fire UAV was tested in simulated high-rise fire scenarios, where flames exhibited rapid size changes and occasional occlusion. The Kalman滤波算法与均值漂移相结合的算法 demonstrated improved prediction speed and reduced target loss compared to standalone mean shift. Performance metrics are summarized in Table 2, highlighting the stability, precision, and anti-interference capabilities. The fire UAV successfully identified and tracked火焰 targets, enabling precise灭火弹 deployment during initial fire stages. Experimental results showed that the algorithm reduced visual recognition failure rates by approximately 30% in高温干扰 conditions, with position prediction accuracy exceeding 90% in controlled tests. These outcomes underscore the effectiveness of the fire UAV system in enhancing灭火救援 for high-rise buildings, where traditional methods often fail due to environmental constraints.
| Metric | Value | Description |
|---|---|---|
| Prediction Stability | High (≥95% consistency) | Measured as the ability to maintain tracking under size changes |
| Position Precision | Within ±0.5 meters | Accuracy of target localization in fire UAV operations |
| Anti-interference Ability | Improved by 40% | Resistance to高温干扰 and occlusion compared to baseline |
| Processing Speed | 5–10 ms per frame | Time for algorithm execution on NVIDIA TX1 platform |
| Target Loss Rate | Reduced to 5% from 15% | Percentage of frames where tracking failed in fire UAV tests |
| Aiming Error | 20 ± 0.5 meters | Overall error in灭火弹 deployment for fire UAV |
The mathematical formulation of the combined algorithm can be extended to optimize parameters for fire UAV applications. Let the target state in the Kalman filter be represented as \(x = [p_x, p_y, v_x, v_y]^T\), where \(p\) denotes position and \(v\) velocity in image coordinates. The state transition matrix \(A\) is defined based on constant velocity motion:
$$A = \begin{bmatrix} 1 & 0 & \Delta t & 0 \\ 0 & 1 & 0 & \Delta t \\ 0 & 0 & 1 & 0 \\ 0 & 0 & 0 & 1 \end{bmatrix}$$
where \(\Delta t\) is the time interval between frames. The observation matrix \(B\) extracts position measurements:
$$B = \begin{bmatrix} 1 & 0 & 0 & 0 \\ 0 & 1 & 0 & 0 \end{bmatrix}$$
For the mean shift part, the target histogram \(q\) is computed from a region of interest (ROI) in the initial frame, with bins based on color features relevant to flames, such as red and yellow intensities. The candidate histogram \(p(y)\) at location \(y\) is compared using the Bhattacharyya distance:
$$\rho(y) = \sum_{u=1}^{m} \sqrt{p_u(y) q_u}$$
where \(m\) is the number of bins. The mean shift vector is calculated as:
$$y_{\text{new}} = \frac{\sum_{i=1}^{n} x_i w_i g\left(\left\|\frac{y – x_i}{h}\right\|^2\right)}{\sum_{i=1}^{n} w_i g\left(\left\|\frac{y – x_i}{h}\right\|^2\right)}$$
where \(x_i\) are pixel locations, \(w_i\) are weights based on histogram similarity, \(g\) is the kernel profile, and \(h\) is the bandwidth. In fire UAV scenarios, I adapt \(h\) dynamically to account for flame size changes, using the Kalman filter’s prediction to adjust the search region. This integration enhances the fire UAV’s ability to track expanding or shrinking火焰 targets, which is common in high-rise fires due to热气流 and fuel variations.
In terms of implementation, the fire UAV software pipeline involves multiple stages. First, image acquisition from onboard cameras captures video feeds at 30 fps. Second, preprocessing steps like noise reduction and color space conversion (e.g., to HSV for better flame detection) are applied. Third, the combined algorithm runs on the AI platform, with the Kalman filter initialized using detections from background subtraction or color thresholding. The fire UAV’s real-time constraints require efficient code; I utilize OpenCV and CUDA加速 on the NVIDIA TX1 to achieve the desired frame rates. The system also incorporates fail-safes, such as reverting to manual control if the algorithm fails, ensuring operational safety. This robust design is essential for fire UAV deployments in unpredictable high-rise environments, where reliability can mean the difference between containment and catastrophe.
The advantages of using fire UAVs with advanced visual algorithms are manifold. Economically, fire UAVs reduce costs compared to manned aircraft or specialized消防 vehicles. Socially, they minimize risks to firefighters by allowing remote operations in hazardous zones. For high-rise buildings, the fire UAV’s ability to reach great heights quickly enables early intervention, potentially preventing fires from escalating into立体火灾. The combined algorithm specifically addresses challenges like GPS denial and visual obstructions, making the fire UAV a versatile tool in urban firefighting. Future work could focus on enhancing the algorithm with deep learning techniques for more robust flame recognition, or on multi-fire UAV coordination for large-scale incidents. Additionally, integrating thermal imaging could improve performance in smoky conditions, further boosting the fire UAV’s effectiveness. As technology advances, I envision fire UAVs becoming standard equipment in fire departments worldwide, transforming rescue operations and saving lives.
In conclusion, the visual algorithm combining Kalman filtering and mean shift offers a significant improvement for fire UAVs in high-rise building firefighting. By addressing target loss due to size changes and occlusion, it enhances the reliability and precision of flame tracking, enabling accurate灭火弹 deployment. The fire UAV system, with its high-temperature resistance and AI-driven control, represents a leap forward in overcoming the limitations of traditional equipment. Experimental results confirm the algorithm’s stability and anti-interference capabilities, meeting the demands of initial fire response. As I continue to refine this approach, the potential for fire UAVs to reduce casualties and economic losses grows, underscoring their value in modern firefighting. The journey toward fully autonomous fire UAV fleets is underway, and this research contributes a critical piece to that puzzle, paving the way for safer and more efficient灭火救援 in our increasingly vertical urban landscapes.
