Abstract
During this era of digital transformation, applications in Logistics and Supply Chain based on Operations Research (OR) and Machine Learning (ML) techniques has catalyzed the development of innovative approaches that redefine industry standards. Particularly, transportation dispatching – a critical aspect of logistics – has seen significant advancements through the application of Reinforcement Learning (RL), achieving notable enhancements in operational efficiency. Despite these advancements, current research predominantly focuses on ride-sharing and on-demand delivery, with limited attention to fair dispatch practices by the workers point of view. This research addresses this gap by proposing a fair truck dispatch system designed to equitably distribute loads from a shipping company to carriers without compromise service levels. Utilizing real-world data characterized by uncertain demand and a dynamic fleet size our empirical results demonstrate the effectiveness of the proposed dispatch strategy, confirming its capability to ensure equitable load distribution across different operational scenarios reaching 86% on average allocation over carriers, compared to their available capacity.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Abahussein, S., Ye, D., Zhu, C., Cheng, Z., Siddique, U., Shen, S.: Multi-agent reinforcement learning for online food delivery with location privacy preservation. Information 14(11), 597 (2023). https://doi.org/10.3390/info14110597
Alibaba. https://damo.alibaba.com/?language=en. Accessed: 27 May 2024
Amazon. https://www.amazon.science/. Accessed: 27 May 2024
ArcelorMittal. https://corporate.arcelormittal.com/about. Accessed: 27 May 2024
ArcelorMittal Brasil. https://brasil.arcelormittal.com/a-arcelormittal/quem-somos. Accessed: 27 May 2024
Brockman, G., et al.: Openai gym. arXiv preprint arXiv:1606.01540 (2016). https://doi.org/10.48550/arXiv.1606.01540
Chen, J., Umrawal, A.K., Lan, T., Aggarwal, V.: DeepFreight: Integrating Deep Reinforcement Learning and Mixed Integer Programming for Multi-transfer Truck Freight Delivery. arXiv preprint arXiv:2103.03450 (2021). https://doi.org/10.48550/arXiv.2103.03450
Chen, Y., et al.:. Can sophisticated dispatching strategy acquired by reinforcement learning? In: Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, AAMAS ‘19, Richland, SC, pp. 1395–1403. International Foundation for Autonomous Agents and Multiagent Systems (2019). https://doi.org/10.48550/arXiv.1903.02716
Dantzig, G.B., Ramser, J.H.: The truck dispatching problem. Manage. Sci. 6(1), 80–91 (1959). https://doi.org/10.1287/mnsc.6.1.80
DiDi. Available at: https://www.didiglobal.com/science/ailabs [Accessed: 27 May 2024]
DoorDash. https://doordash.engineering/. Accessed: 27 May 2024
Liang, E., et al.: RLlib: Abstractions for distributed reinforcement learning. In International conference on machine learning, pp. 3053–3062. PMLR (2018). https://doi.org/10.48550/arXiv.1712.09381
Jiang, L., Wang, S., Guo, B., Wang, H., Zhang, D., Wang, G.: Faircod: A fairness-aware concurrent dispatch system for large-scale instant delivery services. In: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD ‘23, New York, NY, USA, pp. 4229–4238. Association for Computing Machinery (2023). https://doi.org/10.1145/3580305.3599824
Lin, K., Zhao, R., Xu, Z., Zhou, J.: Efficient large-scale fleet management via multi-agent deep reinforcement learning. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ‘18). Association for Computing Machinery, New York, NY, USA, pp. 1774–1783 (2018).. https://doi.org/10.1145/3219819.3219993
Konda, V.R., Tsitsiklis, J.N.: OnActor-Critic Algorithms. SIAM J. Control Optim. 42(4), 1143–1166 (2003). https://doi.org/10.1137/S0363012901385691
Mnih, V., et al.: Asynchronous methods for deep reinforcement learning. In International conference on machine learning, pp. 1928–1937. PMLR (2016)
Mnih, V., et al.:. Playing atari with deep reinforcement learning (2013). arXiv preprint arXiv:1312.5602. https://doi.org/10.48550/arXiv.1312.5602
Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal Policy Optimization Algorithms. ArXiv, abs/1707.06347 (2017). https://doi.org/10.48550/arXiv.1707.06347
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. A Bradford Book, Cambridge, MA, USA (2018)
Uber and UberEats. https://www.uber.com/blog/research/. Accessed: 27 May 2024
Xu, Z., et al.: Large-scale order dispatch in on-demand ride-hailing platforms: A learning and planning approach. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD ‘18, New York, NY, USA, pp. 905–913. Association for Computing Machinery (2018). https://doi.org/10.1145/3219819.3219824
Yan, C., Zhu, H., Korolko, N., Woodard, D.: Dynamic Pricing and Matching in Ride-Hailing Platforms (October 1, 2018). Naval Research Logistics, Forthcoming. https://doi.org/10.2139/ssrn.3258234
Zong, Z., Feng, T., Xia, T., Jin, D., Li, Y.: Deep reinforcement learning for demand driven services in logistics and transportation systems: A survey. CoRR abs/2108.04462 (2021). https://doi.org/10.48550/arXiv.2108.04462
Acknowledgments
The authors wish to express their gratitude to Leandro Galinari, Marlon Lacerda, Talita Bezerra, and Victor Castro of ArcelorMittal Brasil for providing the dataset and their insights into the business processes.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Tavares, M.A.A., Veloso, A.A. (2024). Using Deep Reinforcement Learning to Dispatch Loads to Carriers Under Uncertain Demand and Dynamic Fleet Size. In: Garrido, A., Paternina-Arboleda, C.D., Voß, S. (eds) Computational Logistics. ICCL 2024. Lecture Notes in Computer Science, vol 15168. Springer, Cham. https://doi.org/10.1007/978-3-031-71993-6_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-71993-6_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-71992-9
Online ISBN: 978-3-031-71993-6
eBook Packages: Computer ScienceComputer Science (R0)