Reinforcement learning for humanitarian relief distribution with trucks and UAVs under travel time uncertainty
Heuristics
Time horizon
Vehicle Routing Problem
Robustness
DOI:
10.1016/j.trc.2023.104401
Publication Date:
2023-11-07T23:33:55Z
AUTHORS (3)
ABSTRACT
Effective humanitarian relief operations are challenging in the aftermath of disasters, as trucks often faced with considerable travel time uncertainties due to damaged transportation networks. Efficient deployment Unmanned Aerial Vehicles (UAVs) potentially mitigates this problem, supplementing truck fleets an impactful manner. To plan last-mile distribution setting, we introduce a multi-trip, split-delivery vehicle routing problem and UAVs, soft windows, stochastic times for distribution, formulated dynamic program. Within finite horizon, aim maximize weighted objective function comprising number goods delivered, different locations visited, late arrival penalties. Our study offers insights into dealing uncertainty logistics by (i) deploying partial substitutes trucks, (ii) evaluating solutions generated two deep reinforcement learning (RL) approaches – specifically value approximation (VFA) policy (PFA) (iii) comparing RL stemming from mathematical programming heuristics. Experiments performed on both Solomon-based instances real-world cases. The cases 2015 Nepal earthquake 2018 Indonesia tsunami based locally collected field data UAV specifications, provide practical insights. experimental results show that decision-making improves performance robustness operations, achieving reductions lateness penalties around 85% compared static expected times. Furthermore, replacing half UAVs 11% 56%, benefitting reliability location coverage. indicate use methods successfully mitigate operations.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (95)
CITATIONS (12)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....