Content area

Abstract

Rational location planning of express delivery stations (EDS) is crucial for enhancing the quality and efficiency of urban logistics. The spatial heterogeneity of logistics demand across urban areas highlights the importance of adopting a scientific approach to EDS location planning. To tackle the issue of strategy misalignment caused by heterogeneous demand scenarios, this study proposes a continuous location method for EDS based on multi-agent deep reinforcement learning. The method formulates the location problem as a continuous maximum coverage model and trains multiple agents with diverse policies to enable adaptive decision-making in complex urban environments. A direction-controlled continuous movement mechanism is introduced to facilitate an efficient search and high-precision location planning. Additionally, a perception system based on local observation is designed to rapidly capture heterogeneous environmental features, while a local–global reward feedback mechanism is established to balance localized optimization with overall system benefits. Case studies conducted in Fuzhou, Fujian Province and Shenzhen, Guangdong Province, China, demonstrate that the proposed method significantly outperforms traditional heuristic methods and the single-agent deep reinforcement learning method in terms of both coverage rate and computational efficiency, achieving an increase in population coverage of 9.63 and 15.99 percentage points, respectively. Furthermore, by analyzing the relationship between the number of stations and coverage effectiveness, this study identifies optimal station configuration thresholds for different urban areas. The findings provide a scientific basis for investment decision-making and location planning in EDS construction.

Details

1009240
Business indexing term
Title
A Multi-Agent Deep Reinforcement Learning Method with Diversified Policies for Continuous Location of Express Delivery Stations Under Heterogeneous Scenarios
Author
Lyu Yijie 1 ; Tang Zhongan 2 ; Li, Yalun 1 ; Liu Baoju 3   VIAFID ORCID Logo  ; Deng, Min 3 ; Wu, Guohua 4 

 School of Geosciences and Info-Physics, Central South University, Changsha 410083, China; [email protected] (Y.L.); [email protected] (Y.L.); 
 The Third Surveying and Mapping Institute of Hunan Province, Changsha 410018, China, Hunan Geospatial Information Engineering and Technology Research Center, Changsha 410018, China 
 School of Geosciences and Info-Physics, Central South University, Changsha 410083, China; [email protected] (Y.L.); [email protected] (Y.L.);, The Third Surveying and Mapping Institute of Hunan Province, Changsha 410018, China 
 School of Automation, Central South University, Changsha 410083, China 
Volume
14
Issue
12
First page
461
Number of pages
24
Publication year
2025
Publication date
2025
Publisher
MDPI AG
Place of publication
Basel
Country of publication
Switzerland
Publication subject
e-ISSN
22209964
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Publication history
 
 
Online publication date
2025-11-24
Milestone dates
2025-09-24 (Received); 2025-11-20 (Accepted)
Publication history
 
 
   First posting date
24 Nov 2025
ProQuest document ID
3286304186
Document URL
https://www.proquest.com/scholarly-journals/multi-agent-deep-reinforcement-learning-method/docview/3286304186/se-2?accountid=208611
Copyright
© 2025 by the authors. Published by MDPI on behalf of the International Society for Photogrammetry and Remote Sensing. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2025-12-24
Database
ProQuest One Academic