Content area
What are the main findings? We design a three-layer emergency logistics network to manage the flow of disaster relief materials and develop a bi-objective, multi-period stochastic integer programming model to support post-disaster decision-making under uncertainty. Multi-armed bandit approaches are innovatively applied to solve the problem. A newly developed multi-armed bandit (reinforcement learning) algorithm called the Geometric Greedy algorithm, achieves overall higher performance than the traditional ϵ-Greedy algorithm and the Upper Confidence Bound (UCB) algorithm. What is the implication of the main finding? The key advantage of using reinforcement learning to solve our problem is that agents can dynamically adjust their strategies through interaction with the uncertain environment to minimize action costs. Natural disasters (e.g., floods, earthquakes) significantly impact citizens, economies, and the environment worldwide. Due to their sudden onset, devastating effects, and high uncertainty, it is crucial for emergency departments to take swift action to minimize losses. Among these actions, planning the locations of relief supply distribution centers and dynamically allocating supplies is paramount, as governments must prioritize citizens’ safety and basic living needs following disasters. To address this challenge, this paper develops a three-layer emergency logistics network to manage the flow of emergency materials, from warehouses to transfer stations to disaster sites. A bi-objective, multi-period stochastic integer programming model is proposed to solve the emergency location, distribution, and allocation problem under uncertainty, focusing on three key decisions: transfer station selection, upstream emergency material distribution, and downstream emergency material allocation. We introduce a multi-armed bandit algorithm, named the Geometric Greedy algorithm, to optimize transfer station planning while accounting for subsequent dynamic relief supply distribution and allocation in a stochastic environment. The new algorithm is compared with two widely used multi-armed bandit algorithms: the
Abstract
Details
Emergency medical care;
Integer programming;
Floods;
Greedy algorithms;
Distribution centers;
Machine learning;
Uncertainty;
Climate change;
Efficiency;
Case studies;
Disaster relief;
Emergency preparedness;
Transfer stations;
Planning;
Decision making;
Optimization;
Natural disasters;
Disasters;
Literature reviews;
Algorithms;
Logistics;
Supplies;
Warehouses;
Multi-armed bandit problems
1 Department of Statistics and Data Science, Southern University of Science and Technology, Shenzhen 518055, China;
2 School of Public Administration and Emergency Management, Jinan University, Guangzhou 510632, China
3 Department of Mathematical Sciences, Durham University, Durham DH1 3LE, UK;