强化学习
计算机科学
马尔可夫决策过程
出租车
钥匙(锁)
基线(sea)
过程(计算)
人工智能
组分(热力学)
深度学习
建筑
运筹学
马尔可夫过程
计算机安全
艺术
统计
海洋学
物理
数学
运输工程
工程类
热力学
地质学
操作系统
视觉艺术
作者
Bolong Zheng,Lingfeng Ming,Qi Hu,Zhipeng Lü,Guanfeng Liu,Xiaofang Zhou
摘要
Online ride-hailing platforms have reduced significantly the amounts of the time that taxis are idle and that passengers spend on waiting. As a key component of these platforms, the fleet management problem can be naturally modeled as a Markov Decision Process, which enables us to use the deep reinforcement learning. However, existing studies are proposed based on simplified problem settings that fail to model the complicated supply-dynamics and restrict the performance in the real traffic environment. In this article, we propose a supply-demand-aware deep reinforcement learning algorithm for taxi dispatching, where we use a deep Q-network with action sampling policy, called AS-DQN, to learn an optimal dispatching policy. Furthermore, we utilize a dueling network architecture, called AS-DDQN, to improve the performance of AS-DQN. Extensive experiments on real-world datasets offer insight into the performance of our model and show that it is capable of outperforming the baseline approaches.
科研通智能强力驱动
Strongly Powered by AbleSci AI