Design and Implementation of an Image Acquisition and Recognition Device for High-Voltage Power Line Inspection Using Surveying UAVs

Maintaining the reliability of power transmission infrastructure is critical for industrial operations and public livelihood. Foreign objects on high-voltage overhead lines, such as plastic bags or bird nests, pose significant risks of electrical faults. Traditional inspection methods require manual line walking or post-flight image screening—both labor-intensive and time-consuming. Surveying drones (UAVs) offer an efficient alternative but face limitations in real-time anomaly detection and autonomous navigation without expensive hardware upgrades. This work addresses these gaps by designing a lightweight external device that enables standard surveying UAVs to perform autonomous inspections with embedded real-time foreign object identification.

The system architecture integrates autonomous navigation, real-time image processing, and data classification. As shown in the functional flow below, the device imports predefined flight paths and reference images. During operation, it continuously monitors battery levels, executes autonomous routing, validates captured images against trained models, categorizes results, and transmits alerts. All computations occur onboard using a Raspberry Pi 4B microcomputer, minimizing dependency on ground stations.

Module Components Functionality
Processing Core Raspberry Pi 4B Runs YOLOv5s inference, controls sensors
Positioning AT6558 GNSS chip Provides latitude, longitude, speed, heading
Communication Type-C interface Data/power transfer to UAV
Alert System Programmable buzzer Audible warnings for anomalies

Hardware implementation emphasizes minimal weight (141g) and safety. The polyethylene enclosure attaches beneath the surveying drone via adjustable straps. Power and data exchange occur through a Type-C connector, leveraging the UAV’s battery without modifying internal circuits. Thermal management is achieved through an open-top design and passive cooling. Mounting flexibility allows integration of auxiliary sensors like thermal cameras.

Autonomous navigation uses GNSS data for real-time path correction. The device calculates deviations between planned and actual coordinates, generating steering commands. Positional accuracy follows the AT6558’s 2.5m CEP specification. Flight dynamics are governed by PID control:

$$ \Delta u(t) = K_p e(t) + K_i \int_0^t e(\tau) d\tau + K_d \frac{de(t)}{dt} $$

where \( \Delta u(t) \) represents the control output (throttle/pitch/yaw adjustments), \( e(t) \) is the positional error, and \( K_p \), \( K_i \), \( K_d \) are tuned gains. This enables centimeter-precision hovering during image capture.

Foreign object detection employs YOLOv5s for its optimal speed-accuracy tradeoff on edge devices. The model is trained on a custom dataset of 4,000 annotated images containing common line obstructions. Transfer learning starts from pretrained COCO weights, with hyperparameters configured as:

Parameter Value Description
Epochs 100 Training iterations
Batch Size 4 Limited by Pi’s RAM
Input Size 640×640 Image resolution
nc 1 Single-class (foreign objects)

Training convergence is shown in the mAP and precision curves. After 100 epochs, precision stabilizes at 0.8, while mAP@0.5 reaches 0.5—sufficient for real-time alerts despite dataset limitations. The loss function combines localization, confidence, and classification terms:

$$ \mathcal{L} = \lambda_{\text{coord}} \sum_{i=0}^{S^2} \mathbb{I}_{ij}^{\text{obj}} \left[ (x_i – \hat{x}_i)^2 + (y_i – \hat{y}_i)^2 \right] + \lambda_{\text{coord}} \sum_{i=0}^{S^2} \mathbb{I}_{ij}^{\text{obj}} \left[ (\sqrt{w_i} – \sqrt{\hat{w}_i})^2 + (\sqrt{h_i} – \sqrt{\hat{h}_i})^2 \right] + \sum_{i=0}^{S^2} \mathbb{I}_{ij}^{\text{obj}} (C_i – \hat{C}_i)^2 + \lambda_{\text{noobj}} \sum_{i=0}^{S^2} \mathbb{I}_{ij}^{\text{noobj}} (C_i – \hat{C}_i)^2 + \sum_{i=0}^{S^2} \mathbb{I}_{i}^{\text{obj}} (p_i(c) – \hat{p}_i(c))^2 $$

where \( \mathbb{I}_{ij}^{\text{obj}} \) indicates object presence in grid cell \( i \) and anchor \( j \), \( (x,y,w,h) \) are bounding box coordinates, \( C \) is object confidence, and \( p(c) \) is class probability.

Field tests validated both navigation and detection modules. Autonomous routing trials used a DJI Mini 3 Pro surveying UAV covering 500m of simulated transmission lines. Path-tracking accuracy remained within ±3m despite wind interference, with 7% battery consumed over 2 minutes. For object recognition, black wires suspended plastic bags at 8m height. The surveying UAV identified anomalies at 30°–60° tilt angles and various altitudes (5–15m), achieving 12 FPS inference on Raspberry Pi. Environmental robustness was confirmed across lighting conditions (1,000–80,000 lux) and temperatures (15°C–31°C).

Test Condition Result Performance
Navigation 500m route, 5m/s wind Completed in 120s Positional error < 3m
Detection Plastic bag @ 8m height Real-time ID 12 FPS, 78% precision
Environmental 15°C–31°C, 5–80k lux No image distortion mAP@0.5 = 0.51

This work demonstrates a practical solution for automating high-voltage line inspections using accessible surveying drones. By offloading autonomous navigation and real-time object detection to an external module, the system eliminates post-flight image screening delays. The Type-C interface ensures compatibility with commercial UAVs without internal modifications. Future versions will integrate infrared sensors for thermal anomaly detection and expand the dataset to improve recognition accuracy for rare obstructions. The approach significantly lowers barriers to automated power infrastructure monitoring, particularly in resource-limited settings.

Scroll to Top