Full Text

Turn on search term navigation

© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

This work is focused on the preliminary stage of the 3D drone tracking challenge, namely the precise detection of drones on images obtained from a synchronized multi-camera system. The YOLOv5 deep network with different input resolutions is trained and tested on the basis of real, multimodal data containing synchronized video sequences and precise motion capture data as a ground truth reference. The bounding boxes are determined based on the 3D position and orientation of an asymmetric cross attached to the top of the tracked object with known translation to the object’s center. The arms of the cross are identified by the markers registered by motion capture acquisition. Besides the classical mean average precision (mAP), a measure more adequate in the evaluation of detection performance in 3D tracking is proposed, namely the average distance between the centroids of matched references and detected drones, including false positive and false negative ratios. Moreover, the videos generated in the AirSim simulation platform were taken into account in both the training and testing stages.

Details

Title
YOLOv5 Drone Detection Using Multimodal Data Registered by the Vicon System
Author
Lindenheim-Locher, Wojciech 1 ; Świtoński, Adam 2   VIAFID ORCID Logo  ; Krzeszowski, Tomasz 3   VIAFID ORCID Logo  ; Paleta, Grzegorz 4 ; Hasiec, Piotr 4 ; Josiński, Henryk 2   VIAFID ORCID Logo  ; Paszkuta, Marcin 1   VIAFID ORCID Logo  ; Wojciechowski, Konrad 1   VIAFID ORCID Logo  ; Rosner, Jakub 1   VIAFID ORCID Logo 

 Polish-Japanese Academy of Information Technology, ul. Koszykowa 86, 02-008 Warsaw, Poland; [email protected] (W.L.-L.); [email protected] (G.P.); [email protected] (P.H.); [email protected] (M.P.); [email protected] (K.W.) 
 Department of Computer Graphics, Vision and Digital Systems, Silesian University of Technology, ul. Akademicka 16, 44-100 Gliwice, Poland; [email protected] (A.Ś.); [email protected] (H.J.); Polish-Japanese Academy of Information Technology, ul. Koszykowa 86, 02-008 Warsaw, Poland; [email protected] (W.L.-L.); [email protected] (G.P.); [email protected] (P.H.); [email protected] (M.P.); [email protected] (K.W.) 
 Faculty of Electrical and Computer Engineering, Rzeszow University of Technology, al. Powstancow Warszawy 12, 35-959 Rzeszow, Poland; [email protected]; Polish-Japanese Academy of Information Technology, ul. Koszykowa 86, 02-008 Warsaw, Poland; [email protected] (W.L.-L.); [email protected] (G.P.); [email protected] (P.H.); [email protected] (M.P.); [email protected] (K.W.) 
 Polish-Japanese Academy of Information Technology, ul. Koszykowa 86, 02-008 Warsaw, Poland; [email protected] (W.L.-L.); [email protected] (G.P.); [email protected] (P.H.); [email protected] (M.P.); [email protected] (K.W.); Department of Computer Graphics, Vision and Digital Systems, Silesian University of Technology, ul. Akademicka 16, 44-100 Gliwice, Poland; [email protected] (A.Ś.); [email protected] (H.J.) 
First page
6396
Publication year
2023
Publication date
2023
Publisher
MDPI AG
e-ISSN
14248220
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2843122515
Copyright
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.