Research on UAV Remote Sensing Image Registration Based on Improved SURF Algorithm

In recent years, the rapid advancement of China UAV drone technology has revolutionized remote sensing applications, enabling high-resolution image acquisition for various fields such as environmental monitoring, agriculture, and infrastructure inspection. However, UAV remote sensing images often suffer from significant distortions due to factors like varying shooting angles, illumination changes, and sensor movements, which complicates image analysis and integration. Traditional image registration methods, while developed, frequently fall short in handling these distortions effectively, leading to suboptimal alignment and information loss. To address this, we propose an enhanced image registration method leveraging an improved Speeded-Up Robust Features (SURF) algorithm, tailored for China UAV drone imagery. This approach aims to boost registration accuracy and robustness by refining feature extraction and matching processes, ultimately supporting critical tasks like land management and disaster response in China’s diverse landscapes.

Image registration is a fundamental step in remote sensing, involving the alignment of two or more images of the same scene taken from different viewpoints or times. For China UAV drone operations, this process is crucial for creating seamless mosaics, detecting changes, and enhancing data fusion. The core challenge lies in managing large geometric distortions and noise inherent in drone-captured images. Existing methods, such as those based on curvature scale space or AKAZE algorithms, often struggle with computational efficiency and accuracy under these conditions. Our method introduces improvements to the SURF algorithm, which is known for its speed and robustness, by optimizing keypoint detection and description phases. This not only accelerates processing but also enhances feature matching precision, making it ideal for real-time applications with China UAV drones. In this article, we detail the design, implementation, and validation of our approach, emphasizing its superiority through extensive experiments.

The registration of UAV remote sensing images involves finding a spatial transformation that aligns a reference image with a target image. Mathematically, this can be expressed as: $$ f_{\text{referto}}(x,y) = B(f_{\text{Toberegistered}}(x,y)) $$ where \( f_{\text{Toberegistered}}(x,y) \) and \( f_{\text{referto}}(x,y) \) represent the target and reference images, respectively, and \( B \) is a 2D spatial geometric transformation matrix influenced by factors like illumination and noise. Our method follows a systematic pipeline: image acquisition, preprocessing, feature extraction using improved SURF, similarity measurement, spatial transformation, and coarse-to-fine registration. This structured approach ensures high accuracy even for distorted images from China UAV drone surveys, which often cover complex terrains and urban areas.

Methodology Design

The design of our UAV remote sensing image registration method is grounded in the principles of remote sensing imaging and computer vision. We focus on handling images captured by China UAV drones, which are prone to haze, geometric distortions, and noise due to environmental and operational factors. The overall workflow includes image sample collection, initial preprocessing, feature extraction with improved SURF, similarity度量, spatial transformation determination, and registration execution. Below, we elaborate on each component, incorporating formulas and tables to summarize key aspects.

UAV Remote Sensing Image Acquisition

To obtain UAV remote sensing images, we utilize drones equipped with high-resolution cameras, simulating the operational conditions of China UAV drone fleets. The imaging principle accounts for angles and altitudes, generating samples for both reference and target images. For a given ground point with coordinates \( (x_{\text{surfacefeatures}}, y_{\text{surfacefeatures}}) \), the image pixel is derived as: $$ f_{\text{Toberegistered}}(x,y) = \begin{bmatrix} x_{\text{surfacefeatures}} \cdot \sin \theta_1 \cdot H_1 \cdot \kappa_{\text{remotesense}} \\ y_{\text{surfacefeatures}} \cdot \cos \theta_1 \cdot H_1 \cdot \kappa_{\text{remotesense}} \end{bmatrix} $$ where \( \theta_1 \) and \( H_1 \) are the imaging angle and coefficient for the reference image, and \( \kappa_{\text{remotesense}} \) is the remote sensing imaging coefficient. By adjusting parameters, we collect multiple image pairs, ensuring diversity in scale, rotation, and局部 views to mimic real-world China UAV drone scenarios.

Image Preprocessing

Preprocessing is critical to enhance image quality before registration. We apply three main steps: dehazing, geometric correction, and denoising. First, dehazing addresses atmospheric interference common in China UAV drone imagery. The hazy image model is: $$ f(x,y) = f_0(x,y) \times \eta_{\text{transmission}} + G $$ where \( \eta_{\text{transmission}} \) is the global transmission rate and \( G \) is the atmospheric light vector. The dehazed image is recovered as: $$ f_0(x,y) = \frac{f(x,y) – G}{\eta_{\text{transmission}}} $$ Second, geometric correction reduces distortions via scaling, rotation, and translation. For scaling, the transformation is: $$ f_{\text{check}}(x,y) = \kappa_{\text{Scale}} \cdot f(x,y) $$ where \( \kappa_{\text{Scale}} \) is the scaling factor. Third, denoising employs spatial and frequency domain filters to suppress noise, improving feature extraction reliability. These steps collectively prepare images for subsequent processing, as summarized in Table 1.

Table 1: Summary of Image Preprocessing Steps for China UAV Drone Imagery
Step Purpose Key Formula/Parameter
Dehazing Remove atmospheric haze $$ f_0(x,y) = \frac{f(x,y) – G}{\eta_{\text{transmission}}} $$
Geometric Correction Correct distortions (scale, rotate, translate) $$ f_{\text{check}}(x,y) = \kappa_{\text{Scale}} \cdot f(x,y) $$
Denoising Reduce noise via filtering Low-pass filters in frequency domain

Feature Extraction Using Improved SURF Algorithm

The improved SURF algorithm enhances the standard SURF by optimizing keypoint detection and descriptor computation. SURF relies on Hessian matrix-based interest point detection and integral images for efficiency. For an integral image \( f_{\text{integral}}(x,y) \), defined as: $$ f_{\text{integral}}(x,y) = \sum_{i=1}^{S_{\text{integral}}} f(x_i, y_i) $$ where \( S_{\text{integral}} \) is the integration region, the Hessian matrix at scale \( \delta \) is: $$ H(f_{\text{integral}}) = \begin{bmatrix} C_{xx}(x,y,\delta) & C_{xy}(x,y,\delta) \\ C_{yx}(x,y,\delta) & C_{yy}(x,y,\delta) \end{bmatrix} $$ Here, \( C_{xx}, C_{xy}, C_{yx}, C_{yy} \) are convolutions of Gaussian functions with the integral image. Our improvements include adaptive scale selection and enhanced descriptor dimensionality. We construct a scale space by varying filter sizes and detect keypoints as local maxima. Then, for each keypoint, we assign a dominant direction using a circular neighborhood and compute a 64-dimensional feature vector \( \tau \): $$ \tau = \left( \sum H(f_{\text{integral}})(x), \sum H(f_{\text{integral}})(y), \sum |\varpi_x|, \sum |\varpi_y| \right) $$ where \( \varpi_x \) and \( \varpi_y \) are Gaussian-weighted Haar wavelet responses in horizontal and vertical directions. This process extracts robust features from China UAV drone images, even under varying conditions.

Similarity Measurement

After feature extraction, we measure similarity between reference and target images to identify matching points. The similarity score \( s \) is computed using Euclidean distance between feature vectors \( \tau_{\text{referto}} \) and \( \tau_{\text{Toberegistered}} \). If \( s \) exceeds a threshold \( s_0 \), the features are considered matched; otherwise, they are discarded. This step ensures that only reliable correspondences proceed to transformation estimation, reducing errors in registration for China UAV drone datasets.

Spatial Transformation

Spatial transformation models the geometric relationship between matched images. Common forms include rigid, affine, and projective transformations. For rigid transformation, which involves rotation and translation, the model is: $$ \begin{bmatrix} x_{\text{Toberegistered}} \\ y_{\text{Toberegistered}} \end{bmatrix} = \begin{bmatrix} \Delta x \\ \Delta y \end{bmatrix} + \begin{bmatrix} \cos \varphi & -\sin \varphi \\ \sin \varphi & \cos \varphi \end{bmatrix} \begin{bmatrix} x_{\text{referto}} \\ y_{\text{referto}} \end{bmatrix} $$ where \( \Delta x \) and \( \Delta y \) are translation offsets, and \( \varphi \) is the rotation angle. We select the transformation type based on feature geometry, ensuring accurate alignment for China UAV drone images with complex distortions.

Registration Implementation: Coarse-to-Fine Approach

Registration is performed in two stages: coarse and fine. Coarse registration uses feature matching to estimate an initial alignment. For matched points, the coarse parameters are derived as: $$ \begin{cases} x_{\text{coarse registration}} = l_x \pm x_{\text{referto}}^{(i)} \\ y_{\text{coarse registration}} = l_y \pm y_{\text{referto}}^{(i)} \end{cases} $$ where \( l_x \) and \( l_y \) are offset components, and \( (x_{\text{referto}}^{(i)}, y_{\text{referto}}^{(i)}) \) is a reference point. Fine registration then applies interpolation to refine alignment. For linear interpolation between points \( P_1(x_1, y_1) \) and \( P_3(x_2, y_1) \), the horizontal interpolation is: $$ x_{\text{interpolation}} = \frac{x_2 – x}{x_2 – x_1} y_{\text{interpolation}}(P_1) + \frac{x – x_1}{x_2 – x_1} y_{\text{interpolation}}(P_3) $$ This coarse-to-fine strategy enhances precision, crucial for China UAV drone applications where high accuracy is demanded.

Experimental Analysis

To evaluate our method, we conducted extensive experiments using UAV remote sensing images from China UAV drone operations. The setup involved hardware and software configurations, dataset preparation, and comparison with traditional methods. Our goal was to assess registration effectiveness through quantitative metrics, highlighting the advantages of the improved SURF algorithm in real-world scenarios.

Experimental Setup

We used a quadcopter drone (Mavic model) equipped with a CMOS camera (Altum model) to capture images over a diverse Chinese region covering approximately 21,454 km². This China UAV drone setup simulates typical survey conditions. Images were collected under varying angles and lighting, then processed into reference and target samples through cropping, rotation, and scaling. The hardware environment included a PC with Intel Xeon E5345 CPU and 2 GB RAM, running MATLAB R2008b for algorithm implementation. Parameters for the improved SURF algorithm were set as: scale space levels = 4, feature point threshold = 1000, and iteration count = 10. We compared our method with two traditional approaches: one based on improved curvature scale space (Method 1) and another on improved AKAZE (Method 2), focusing on registration accuracy and efficiency.

Results and Discussion

We tested our method on three image types: local (cropped), rotated, and scaled, representing common challenges in China UAV drone imagery. Registration performance was measured using peak signal-to-noise ratio (PSNR) and correct registration rate (CRR). PSNR is defined as: $$ \mu = 10 \cdot \lg\left(\frac{255^2}{\varepsilon}\right) $$ where \( \varepsilon \) is the mean square error, and CRR is: $$ \eta_{\text{Registration}} = \frac{N_{\text{correct}}}{N_{\text{characteristic}}} \times 100\% $$ with \( N_{\text{correct}} \) as correctly registered feature points and \( N_{\text{characteristic}} \) as total feature points. Higher values indicate better registration.

For local images, Table 2 presents detailed results across six samples. Our method consistently achieved higher PSNR and CRR than traditional methods, demonstrating superior alignment for China UAV drone data.

Table 2: Registration Results for Local UAV Remote Sensing Images from China UAV Drone
Image Sample Total Feature Points Method 1: Correct Points / PSNR (dB) Method 2: Correct Points / PSNR (dB) Our Method: Correct Points / PSNR (dB)
J01 154 145 / 89.5 150 / 92.1 152 / 114.6
J02 296 280 / 87.1 291 / 95.6 295 / 123.8
J03 650 632 / 85.6 642 / 90.4 647 / 151.3
J04 474 455 / 82.2 466 / 91.5 471 / 109.4
J05 358 331 / 80.7 350 / 96.4 356 / 106.2
J06 410 387 / 89.9 401 / 93.2 408 / 147.9
Average 94.8% CRR / 85.8 dB PSNR 98.1% CRR / 93.2 dB PSNR 99.4% CRR / 125.5 dB PSNR

For rotated images, the PSNR comparison is illustrated in Figure 1 (simulated as a table due to text constraints). Our method maintained high values across rotation angles, with an average CRR of 99.4%, outperforming Method 1 (94.1%) and Method 2 (97.8%). This robustness is vital for China UAV drone applications where images are often captured from different orientations.

Table 3: PSNR (dB) for Rotated UAV Images from China UAV Drone
Rotation Angle (degrees) Method 1 PSNR Method 2 PSNR Our Method PSNR
30 84.2 88.5 120.3
60 82.7 90.1 118.9
90 81.5 89.3 122.4
120 83.9 91.2 119.7
Average 83.1 89.8 120.3

For scaled images, Table 4 summarizes results. Our method achieved an average PSNR of 128.4 dB and CRR of 99.4%, significantly higher than Method 1 (75.8 dB, 93.8%) and Method 2 (78.6 dB, 95.6%). This highlights the effectiveness of the improved SURF algorithm in handling scale variations common in China UAV drone surveys.

Table 4: Registration Performance for Scaled UAV Images from China UAV Drone
Scale Factor Method 1: PSNR (dB) / CRR (%) Method 2: PSNR (dB) / CRR (%) Our Method: PSNR (dB) / CRR (%)
0.5x 74.3 / 92.5 77.2 / 94.8 126.8 / 99.2
1.5x 76.1 / 94.1 78.9 / 95.9 129.1 / 99.5
2.0x 77.0 / 94.8 79.7 / 96.1 129.3 / 99.6
Average 75.8 / 93.8 78.6 / 95.6 128.4 / 99.4

The experimental results consistently show that our method improves PSNR by approximately 50% and CRR by 3.7% on average compared to traditional methods. This enhancement stems from the refined feature extraction and matching in the improved SURF algorithm, which better handles distortions in China UAV drone imagery. The coarse-to-fine registration further boosts accuracy, making it suitable for demanding applications like environmental monitoring and urban planning across China.

Conclusion

In this research, we developed an advanced image registration method based on an improved SURF algorithm, specifically designed for UAV remote sensing images from China UAV drone operations. By integrating enhanced preprocessing, robust feature extraction, and a coarse-to-fine registration strategy, our approach effectively addresses large distortions and noise challenges. Experimental validation using diverse image samples—local, rotated, and scaled—demonstrated significant improvements in peak signal-to-noise ratio and correct registration rate over traditional methods. These findings underscore the method’s potential to enhance data integration and analysis in critical fields such as agriculture, disaster management, and infrastructure inspection within China. Future work will explore applications to flipped and deformed images, further expanding the utility of China UAV drone technology in remote sensing.

Scroll to Top