Content area

Abstract

With the swift evolution of artificial intelligence in the automotive sector, autonomous driving has ascended as a pivotal research frontier for automotive manufacturers. Environmental perception, as the cornerstone of autonomous driving, necessitates innovative solutions to address the intricate challenges posed by data sensitivity during vehicle operations. To this end, federated learning (FL) emerges as a promising paradigm, offering a balance between data privacy preservation and performance optimization for perception tasks. In this paper, we pioneer the integration of FL into 3D object detection, presenting personalized FedM2former, a novel multi-modal framework tailored for autonomous driving. This framework aims to elevate the accuracy and robustness of 3D object detection while mitigating concerns over data sensitivity. Recognizing the heterogeneity inherent in user data, we introduce a personalization strategy leveraging stochastic gradient descent optimization prior to local training, ensuring the global model’s adaptability and generalization across diverse user vehicles. Furthermore, to address the sparsity of point cloud data, we innovate the attention layer within our detection model. Our balanced window attention mechanism innovatively processes both point cloud and image data in parallel within each window, significantly enhancing model efficiency and performance. Extensive experiments on benchmark datasets, including nuScenes, ONCE, and Waymo, demonstrate the efficacy of our approach. Notably, we achieve state-of-the-art results with test mAP and NDS of 71.2% and 73.6% on nuScenes, 67.14% test mAP on ONCE, and 83.9% test mAP and 81.8% test mAPH on Waymo, respectively. These outcomes underscore the feasibility of our method in enhancing object detection performance and speed while safeguarding privacy and data security, positioning Personalized FedM2former as a significant advancement in the autonomous driving landscape.

Details

1009240
Title
Personalized FedM2former: An Innovative Approach Towards Federated Multi-Modal 3D Object Detection for Autonomous Driving
Author
Zhao, Liang 1 ; Li, Xuan 2 ; Jia, Xin 2 ; Fu, Lulu 2 

 College of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China; [email protected] (X.L.); [email protected] (X.J.); [email protected] (L.F.); Henan Key Laboratory of Grain Storage Information Intelligent Perception and Decision Making, Henan University of Technology, Zhengzhou 450001, China 
 College of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China; [email protected] (X.L.); [email protected] (X.J.); [email protected] (L.F.) 
Publication title
Processes; Basel
Volume
13
Issue
2
First page
449
Publication year
2025
Publication date
2025
Publisher
MDPI AG
Place of publication
Basel
Country of publication
Switzerland
Publication subject
e-ISSN
22279717
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Publication history
 
 
Online publication date
2025-02-07
Milestone dates
2025-01-20 (Received); 2025-02-06 (Accepted)
Publication history
 
 
   First posting date
07 Feb 2025
ProQuest document ID
3171217700
Document URL
https://www.proquest.com/scholarly-journals/personalized-fedm-sup-2-former-innovative/docview/3171217700/se-2?accountid=208611
Copyright
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2025-12-10
Database
ProQuest One Academic