Cross-Modal Object Detection via UAV
Document Type
Article
Publication Title
IEEE Transactions on Vehicular Technology
Abstract
UAV-based object detection aims at locating and recognizing targets in aerial images, which is widely applied to traffic surveillance, disaster rescue and anomaly monitoring. However, due to expensive sensors and complicated architectures, it is unrealistic to deploy precise but heavy multi-modal object detectors into UAV nodes. To get over the dilemma, inspired by model compression and cross-modal signal processing techniques, this paper proposes a cross-modal knowledge distillation (CKD) enabled object detection paradigm, which achieves comparable detection performance with multi-modal techniques, yet requires less computational resource. On the one hand, in order to avoid transferring redundant feature knowledge, we design a Selective Feature Imitation (SFI) to selectively shorten the distance between cross-modal features according to their types. On the other hand, in order to transfer the most valuable prediction knowledge, we design an Adaptive Prediction Imitation (API). It evaluates the quality of prediction knowledge, and then adaptively adjusts the distillation intensity for cross-modal prediction. Extensive experiments on the DroneVehicle dataset have shown the performance improvement of the proposed scheme.
First Page
1
Last Page
12
DOI
10.1109/TVT.2023.3262129
Publication Date
3-27-2023
Keywords
Autonomous aerial vehicles, Cameras, Computational modeling, Cross Modality, Detectors, Feature extraction, Knowledge Distillation, Object Detection, Object detection, Training, Unmanned Aerial Vehicles
Recommended Citation
A. Li et al., "Cross-Modal Object Detection via UAV," in IEEE Transactions on Vehicular Technology, pp.1-12, Mar 2023, doi: 10.1109/TVT.2023.3262129.
Comments
IR Deposit conditions:
OA version (pathway a) Accepted version
No embargo
When accepted for publication, set statement to accompany deposit (see policy)
Must link to publisher version with DOI
Publisher copyright and source must be acknowledged