The order dispatching and fleet management system at a ride-sharing company must make decisions both for assigning available drivers to nearby passengers (hereby called orders) and for repositioning drivers who have no nearby orders. These decisions have short-term effects on the revenue generated by the drivers and driver availability. In the long term they change the distribution of drivers across the city, which in turn has a critical impact on how well future orders can be served. Provident algorithmic solutions, which account for the short term and long-term consequences of their decisions can improve the quality of service of the ride-sharing platforms and are thus an important area of research.
Recent works [1, 15] have successfully applied Deep Reinforcement Learning (RL) techniques to dispatching problems, such as the Traveling Salesman Problem (TSP) and the more general Vehicle Routing Problem (VRP) , however they have primarily focused on static (i.e. those where all orders are known up front) and/or single-driver dispatching problems. In contrast to these problems, the fleet management and order dispatching problem ride-sharing platforms face has multiple drivers and dynamically changing supply and demand conditions. We refer to this dynamic dispatching and fleet management problem as the Multi-Driver Vehicle Dispatching and Repositioning Problem (MDVDRP).
VRPs and other problems similar to the MDVDRP are studied in the field of combinatorial optimization. Exactly solving instances of these problems at the scale of real-world environment is computationally intractable. To deal with the intractability, heuristic solutions, which produce approximate solutions in polynomial time, are often used [4, 17]. The planning problem presented by the MDVDRP is related to the VRPs, but the complexity comes from the dynamic nature of the assignment scenario rather than the intractability of computing the exact solution. Drivers and orders appear in the assignment system at random points in time. In this dynamic assignment setting, assignment decisions are made based on the current driver-order situation, without exact information about future orders. A high performing assignment solution needs to account for unknown future supply and demand conditions.
In realistic instances of the MDVDRP, the decision-making continues 24 hours a day and may involve thousands of drivers and tens of thousands of customers. Accounting for the spatially and temporally varying supply and demand conditions makes hand-crafting heuristic solutions to these scenarios challenging. In this paper, we explore a deep reinforcement learning approach to the MDVDRP. The contributions of our work can be summarized as follows. (i) We introduce a new reinforcement learning problem: the MDVDRP, which is motivated by the needs of real-world ride-sharing platforms. (ii) We propose a novel network architecture for decision-making in a realistic problem setting with variable sized observation and action spaces. (iii) We provide empirical analysis of value-based and actor-critic methods on instances of the MDVDRP, including instances based on real-world data.
Ii Related Work
Recent machine learning approaches to dispatching and routing problems operate according to an encoding-decoding scheme, where information is first processed into a fixed-sized representation, and then actions are decoded from this representation [1, 7]. Machine learning approaches to the more general problem family of combinatorial optimization problems are surveyed in . Pointer networks  offer an approximate solution to traveling salesman problems by encoding cities (in our terminology, orders) with a recurrent network, and then producing a solution by sequentially “pointing” to orders using an attention mechanism 
. The network is trained with a supervised loss function by imposing a fixed ordering rule on decoded orders. Belloet al. propose training an architecture similar to the pointer networks with policy gradients instead of a supervised loss. Using policy gradients allows them to dispense with the fixed ordering of the outputs during the decoding phase . Similarly, we use reinforcement learning to train our networks. An alternative to pointer networks is explored in , where graph convolutional networks are used for solving routing problems. We follow an architecture related to , which uses an attention mechanism for encoding the inputs. We depart from their architecture in two ways. First, we replace the input attention layers with layers that compute their output elementwise. Second, we remove the recurrent network used in the decoder.
In practice, order dispatching and fleet management problems are often solved with heuristic solutions, which construct an approximation to the true problem by ignoring the spatial extent, or the temporal dynamics, or both, and solve the approximate problem exactly. One example of a heuristic solution to the order dispatching problem is the myopic pickup distance minimization (MPDM), which ignores temporal dynamics and always assigns the closest available driver to a requesting order . In Local Policy Improvement , handcrafted heuristics are combined with a machine learning method by summarizing supply and demand patterns into a table and then using the learned patterns to account for future gains in the real-time planner of the dispatching solution. A fully machine learning based approach to the dispatching problem is presented in , where deep Q-learning is used for learning dispatching strategies from the perspective of a single driver. We take these developments a step further and learn fleet management and order dispatching strategies end-to-end using reinforcement learning.
Another thread of related work comes from the multi-agent reinforcement learning literature. Specifically, our single-driver training approach is analogous to the “independent Q-learning” training approach . Independent Q-learning has been leveraged successfully in a number of problems including resource allocation problem for elevator control , traffic signal control , and fleet management . Multi-agent reinforcement learning has also been investigated for order dispatching in , where a grid-based algorithm is used. Instead of operating on a grid, our method is based on continuous coordinates, which makes it more ready for deployment in the real-world.
Iii-a MDVDRP as a Reinforcement Learning Environment
The experiments presented in this paper are conducted in a ride-sharing simulation environment. The environment is designed to capture the dynamic supply and demand situation in ride-sharing platforms. The environment represents customers as orders, which start and end in some coordinates. Assigning a driver to the order immediately yields a reward proportional to the price of the order. Drivers are represented as points in the 2-D space and they can move by serving orders or repositioning without an order. The drivers and orders are generated following a Poisson process with parameters depending on each scenario we consider.
Decision points are triggered whenever a driver becomes active in the system, finishes an order or finishes a repositioning action. Variable amounts of time may have passed between the decision points when the simulator polls the policy for actions. The simulation scenarios considered in the experiments mostly depict the situation where there are more customers demanding rides than there are drivers to serve them. In this setting, it is natural to assume that only a single driver is available for actions at any decision point. This assumption simplifies the design of the policies by removing the need to choose which driver to assign an order to. To prevent multiple drivers polling for actions at exactly the same moment, we add random noise with small variance to the duration of the reposition actions. The actions available to the RL agent at each decision point are with respect to the driver that is currently available in the environment. These actions include assigning the driver to nearby orders as well as actions to reposition the driver to other areas.
Assigning a driver to an order removes the order from the system and makes the driver unavailable for instructions until the order has been completed. At order completion the driver is relocated to the order destination and made available for assignments. Orders have a limited time window within which they are valid. After the validity window has passed, the orders will be removed from the system.
The policies may choose to move the drivers to different directions for a fixed amount of time by selecting reposition actions. The number of available reposition actions includes actions that move the driver into one of the eight cardinal directions for a fixed period of time and a stationary action.
The observation of the agents consists of the environment time, the driver for which action is currently being selected and all drivers and orders currently in the simulation. At time there is a collection of orders , drivers , with exactly one available driver
. The state is given to the neural network as
. The orders are presented as 6 dimensional vectors consisting of the starting and ending coordinates, price, and time waiting. Time waiting is the difference between the current time and the creation time. A driver is represented by a 6 dimensional vector: the coordinates of the driver location,and components of its reposition direction, time to order completion, and time to reposition completion. If the driver is serving an order, its location is set to the ending location of the order it is servicing. If the driver is repositioning the driver location is updated at each timestep, the reposition direction shows which way the driver will move during the next timestep.
To limit the number of orders considered by the policy at each timestep, we impose a broadcasting radius on the order assignment. This means that drivers may be only paired with orders if they are within units of the driver. Otherwise, the driver may only take a repositioning action. The repositioning actions are not available to the drivers when there are orders within broadcasting radius of them.
Iii-B Reward Settings
The objective of the algorithms in the MDVDRP problem is to maximize the cumulative reward defined by the environment. The environment rewards the agent for each assigned order with a reward the size of the order price. Reposition actions yield no reward. We consider two alternative reward specifications corresponding to driver-centric and system-centric perspectives. An visual overview of these concepts is presented in Figure 1
In the driver-centric approach, we consider the MDVDRP a reinforcement learning problem from the perspective of the individual driver. In this setting, each driver is maximizing their own expected revenue and there are no incentives for co-operation. The trajectories taken by each driver in the environment are collected separately and the discounted returns are computed on the individual trajectories. The individual trajectories consist of timesteps that are consecutive from the perspective of the driver but not necessarily from the perspective of the environment as other drivers may have taken actions between the actions of any single driver.
From the point of view of the ride-sharing platform, the dispatching and repositioning problem is not about individual drivers maximizing their own revenue but rather about the platform maximizing the combined revenue across all drivers. Therefore, it is important to consider optimizing the policies from the perspective of the whole system. In this system-centric approach, the expected cumulative reward across all drivers is being maximized. This leads to the trajectories experienced by the policy consisting of timesteps that are consecutive from the perspective of the environment. For example, if the driver acts on the timestep receiving reward and driver takes an action on the timestep , the training algorithm will consider the timesteps and consecutive.
Iii-C Neural Network Architecture
We propose a neural network architecture for RL in environments with variable sized observation and action spaces. An overview of the proposed network architecture is presented in Figure 2. A learned pooling mechanism allows the network to compute a fixed-sized global representation of the inputs, which enables relating the features of each individual input to the global state. In the environments we consider, the number of actions depends on the number of orders in the observation as described in III-A. We compute network outputs, one for each action, in a manner similar to the weight computation in the attention mechanism .
The network first computes order embeddings , order pooling weights , driver embeddings , and driver pooling weights . The embeddings are length 128 vectors computed by and
, both of which have one hidden layer of size 128 and ReLU activations. The scalar pooling weightsare computed from the embeddings by and . Both have hidden layer size 128 and tanh activation, and sigmoid output activation. The global context vector is computed as , where denotes concatenation and denotes time.
, which has one hidden layer of size 64 and ReLU activation, outputs values for the assignment actions. takes each order embedding concatenated with the global context separately as input and produces a scalar output for each input. The repositioning actions are computed by with the same architecture as . has one output for each reposition action and uses the global context as its input.
Iii-D Training Algorithms
We train the proposed network architecture on the MDVDRP using two modern reinforcement learning algorithms: Deep Q-Networks (DQN)  and Proximal Policy Optimization (PPO) . We chose these two algorithms because they represent widely used, modern methods in off-policy and on-policy reinforcement learning and many of the recent successes in deep reinforcement learning have been achieved using one of the two algorithms. A short description of the key ideas of the algorithms is given in the following. For more details we refer the reader to the corresponding papers.
DQN is a value-based, off-policy reinforcement learning algorithm. The Q-network estimates state-action values meaning it gives numeric estimates to the expected future return for each action available in the current state. On each iteration, it optimizes the objective function
where are states, actions, rewards and next states sampled from the environment, is a discount factor, is a neural network parameterized by approximating the action-value function, are the target network parameters and
are the learning network parameters. The DQN algorithm collects experience from the environment, stores the experience in a replay buffer and updates the learning network parameters by performing stochastic gradient descent on the loss function. The target network is used to compute an action-value estimate in the next state. In order to improve the stability of the learning network update, the target network parameters are copied from the learning network every couple of hundred update steps. Actions are selected by taking the
of the Q-network output, except with probabilitya random action is chosen in order to make the algorithm explore. When evaluating the policies exploration is no longer needed and the epsilon can be set to a zero.
PPO is an on-policy actor-critic reinforcement learning algorithm . Unlike DQN, PPO directly optimizes the policy using the policy gradient method. In our case, the parameters of the policy are estimated by a neural network. During training time the actions are chosen by sampling from the categorical distribution parameterized by the policy. At evaluation time, we select the actions with the maximum probability. The policy gradient objective is given by
where is the action distribution defined by the policy, are the parameters of the policy, are trajectories sampled from the environment, and is the advantage function. We estimate the advantage function using the generalized advantage estimation method , which requires requires computing a state-value function estimate called the critic. The critic is learned alongside the policy using the same data. On-policy reinforcement learning algorithms require new samples for each update they compute, otherwise they risk deviating too far from the previous policy, which may harm the policy performance. Since sample collection can be expensive in reinforcement learning, updating on each batch of data multiple times would be beneficial. In order to enable multiple updates on the same data, different variants of the PPO algorithm either clip or penalize the objective in order to prevent the policy from diverging too far from the previous one. We use the clipping variant, which clips the objective based on the probability ratio between the current and the previous policy to remove the incentive to diverge too far.
As the simulation runs in continuous time, the time between consecutive timesteps may vary. Both algorithms use a discounting factor to compute the discounted return, which accounts for the future rewards. In continuous time settings, the discounting factor for each timestep is , where is the time at the current observation and in the next.
We found empirically that training using the system-centric rewards is more challenging than using the driver-centric rewards. We found the use of n-step Q-learning  helpful for stabilizing the training of DQN when training on system-centric rewards. In all of our DQN experiments with system-centric reward we use 20-step Q-learning.
|(a) Regional Domain||(b) Hot-Cold Domain|
|(c) Distribute Domain||(d) Historical Orders Domain|
|Hot Cold||Regional||Historical Orders|
|Algorithm||High Demand||Low Demand||High Demand||Low Demand|
|Algorithm||50/50 Served %||80/20 Served %|
We investigate the learning behavior of the proposed network architecture in ride-sharing environments implemented using the simulator described in III-A. We present results in five environments to test and illustrate the dispatching and repositioning strategies learned the proposed approach. The environments are illustrated in Figures 3, 4. The geography of the environments is presented as a rectangle with the longer side length set to one. The drivers move at speed 0.1 and the broadcasting radius is set to 0.3.
We compare the learned policies against two kinds of baselines: myopic revenue maximization (MRM) and myopic pickup distance minimization (MPDM) . MRM always assigns the highest value order to the closest available driver. MPDM assigns orders to drivers in the order of shortest distance first. Since the baselines do not account for repositioning, we test three variants of the baselines with different repositioning heuristics. In the simple variants, broadcasting distance is ignored and drivers can be assigned to orders across the region. The drivers stay where they drop orders off until they are assigned to the next order. In the random variants of MRM and MPDM, if a driver has no orders the within broadcast distance, then a repositioning action is selected randomly. The demand variants of MRM and MPDM apply a simple heuristic on repositioning, by moving the drivers towards the nearest order when an order is available but not within broadcasting distance. If there is no order available, the demand variants take random reposition actions.
We use the same neural network architecture for all variants of the learning algorithms, except for the extra output layer implementing the critic for PPO, which does not exist in the model used for DQN. We set the discount factor for both algorithms. For DQN we use batch size of , replay buffer size and update on every timestep with learning rate . The exploration rate starts from and we anneal it by on each episode to the final value . We update the target network every steps. The PPO policy and value functions are updated
times every epoch usingconsecutive timesteps as the training data. We use generalized advantage estimation (GAE) 
and set the GAE hyperparameter. We aim to keep hyperparameters unchanged between the different environments but make slight changes for Distribute Domain and the real data based domains. Those changes are explained in the relevant sections. To handle the large numbers of drivers and orders in the Historical Statistics domain, we use a parallel implementation of PPO and suitable hyperparameters, which are explained in IV-E.
Iv-a Regional Domain
In our first experiment, we consider the Regional Domain, which illustrates how a simple price differential can be exploited by learned policies but is missed by the myopic dispatching and repositioning approaches. Intuitively, in this domain, a good policy dedicates enough drivers to fully serve the high-reward orders and serves the other orders with the remaining drivers. Myopic policies, which ignore the long-term effects of their decisions, will fail to exploit the available extra reward from serving the high-reward orders. The high-level concept of Regional Domain is presented in Figure 3. In the Regional Domain, there are three regions: left, center, and right. One quarter of all orders go from the center to the upper-left region, one quarter from center to bottom-right, one quarter from upper-left to center, and one quarter from bottom-right to center. All orders yield a reward of 2 except those that go from right to center, which yield a reward of 4.
In the high demand version of the Regional Domain, the proposed method outperforms the baselines by over a 15% margin between the best performing baseline (MPDM-demand) and the best learned policy (Driver-Centric DQN). To identify the source of the difference between the learned policies and the baselines, we look at a policy trained with PPO on driver-centric rewards. The learned policy serves 70 times more orders from the high reward region than the low reward region and achieves significantly higher score than the baselines, which suggests that it has learned to exploit the higher priced orders to its advantage.
The low demand variant proves to be challenging for the learning algorithms. Here the demand is low enough that over commitment to the high reward area can become a problem as is the case for Driver-Centric PPO, which ends up allocating 8 times more idle drivers to the high reward area and as a consequence has lower ratio of total orders served compared to MPDM-demand. This policy leads to weak performance, and is in fact outperformed by the simple and demand variants of MPDM.
Iv-B Hot-Cold Domain
The Hot-Cold domain can be thought of as a ride-sharing scenario with a busy area of downtown (“hot region”) surrounded by suburbs with less traffic (“cold region”). In the Regional domain, the advantage of traveling to the region with higher prices is clear; it is directly tied to the price of orders found in that region (4 vs. 2). In the Hot-Cold domain, the agent must learn a more subtle advantage. Order pickup locations are located uniformly along the top edge of the simulation area. Half of the orders end uniformly along the bottom edge of the area and half end uniformly in the hot region. Order price is given by the Euclidean distance from order pickup to order drop-off locations. Despite orders to the cold region having higher price (since they are longer on average), it is generally more advantageous for drivers to stay in the hot region, since there they can quickly pick up new orders. In other words, the advantage is entirely temporal. An illustration of the Hot-Cold domain is presented in Figure 3.
The results in Table I suggest that learning to balance serving the orders to hot and cold regions is straightforward for the learning algorithms and all learning algorithms outperform all baselines in the high demand variant of the environment. As with the Regional domain, the challenges of training DQN with system-centric rewards are apparent in the low demand Hot-Cold domain, where it is outperformed by multiple baselines.
The margin in favor of the learned policies is narrower in the low demand variant. In the low demand variant of the environment, the policies trained with driver centric rewards the ratio of orders served is approximately double compared to the high demand variant (50% vs 25%). The higher served ratio suggests that there is less freedom of choice between hot-to-hot and hot-to-cold orders in the low demand regime, which in turn reduces the potential for improvement over the naive policies.
In all of the experiment configurations for Regional domain and Hot-Cold domain, we found DQN to be more sample efficient than PPO. DQN was trained for 1000 episodes and PPO often took two to three times as many to achieve similar performance. Actor-critic methods are known to be less sample efficient than value-based methods and another contributing factor may be the relatively low learning rate we used for PPO.
Iv-C Distribute Domain
While Hot-Cold domain tests an important aspect of learning - namely, the ability of the agents to reposition drivers to locations where they can pick up new orders, this repositioning behavior is quite simple in that it is uniform across drivers. This means that each individual driver can always be repositioned in the same manner (i.e. ”if in cold region, go to hot region”). In order to test whether our methods can learn non-uniform repositioning behavior, we introduce a class of “distribution environments” where drivers must be repositioned so as to match their spatial distribution with a fixed future order distribution. A Distribute domain operates in two phases. In the first phase, the environment resets to a state with drivers and no orders, so drivers may only reposition during this phase. In the second phase, orders appear according to a fixed spatial distribution, and drivers can match to orders if they are within a given broadcast radius . The second phase only lasts long enough to allow drivers to reposition one more time before all orders cancel and the environment is reset. Each order matching action receives reward. Order destinations are placed away from start locations so that each driver may only serve one order per episode. As a result, the episodic return is proportional to the number of orders served, so we may interpret the episode score as a measure of how well the agent arranges driver supply in phase 1 with order demand in phase 2.
In our experiments, the distribution of orders always consists of two small patches in the top left and bottom right parts of the unit square, refer to Figure 3 for visualization. The order start locations are sampled uniformly within each patch. The total number of orders in each patch is fixed across episodes, and we denote it fractionally. An even order split between patches (e.g. 10 orders in both patches) is denoted . If 80 percent of orders are in the first patch and 20 percent are in the second patch, we denote it as .
We found that training successful policies on the Distribute Domains requires more random exploration than any of the other environments. For DQN, we set the final . For PPO, we follow an annealing scheme similar to DQN where we anneal the entropy coefficient from to over the first 2000 epochs.
Results for distribute domains with 20 drivers are presented in Table II. We include the optimal served percentage (which is 100 %) and the “uniform optimal” served percentage. This quantity reflects the maximum score one can obtain if the repositioning behavior is uniform across drivers. The results between driver and system-centric variants are mixed. While system-centric PPO achieves a slight advantage over the driver-centric variant, the relationship is flipped for DQN.
Iv-D Historical Orders Domain
Solving the MDVDRP is motivated by the real-world need of ride-sharing platforms to improve their marketplace for drivers and passengers. To test our learning algorithms in a more realistic setting, we consider an environment where the order generation scheme is based on historical order data from the GAIA dataset . The dataset provides gps records for all orders served during a 30-day period in the city of Chengdu in China. To limit computational demands, we choose a subset of consisting of of the orders in the dataset. The orders in the subset all start and end in an area from the dataset that has approximately twenty thousand orders per day. We then create 30 order generation schemes, which correspond to the 30 days in the dataset. Each episode in the environment corresponds to a randomly sampled day from the dataset. The orders generated during the episode appear exactly in the coordinates and at the time defined by the data. We use a fixed number of drivers (100), a 2 kilometre broadcast radius and a fixed speed (40 km/h). Histograms of the order start locations in the area under consideration are presented in Figure 4.
Similarly to the Hot-Cold domain, the reward for serving an order is the distance from the order start location to the drop-off location. Following from the reward definition, the total reward each driver receives is directly correlated with the total time they spend serving orders and thus inversely correlated with the time spend between orders.
Driver-centric DQN and PPO are both capable of learning strong dispatching and repositioning policies, which outperform all of the baselines by over 6%. The learning curves for the experiments in this domain are presented in Figure 7. The historical data domain also demonstrates the importance of redistributing drivers during their downtime. The worst performing algorithms on the historical data domain are the simple variants of MPDM and MRM, which do not use reposition actions at all. In contrast, the learned policies actively redistribute drivers back to the populated areas as shown in Figure Figure 3 (d).
Finally, we observe that both DQN and PPO have trouble learning competitive policies on this larger scale problem when trained on the system-centric rewards. We hypothesize that this is due to having more drivers complicating the already challenging credit assignment problem. With more drivers simultaneously acting in the environment, the time between decision points becomes smaller and the effective discount factor approaches one due to the way it is computed. Training with a discount factor close to one can make the training unstable. To mitigate this effect we use a smaller discount factor .
|(a) Active Hours||(b) Quiet Hours|
Iv-E Historical Statistics Domain
Finally, we test our method in another MDVDRP instance based on the GAIA dataset. In the Historical Statistics domain, both drivers and orders are generated from a Poisson process. Having the number of drivers vary with the time of the day takes the simulation another step closer to the realistic setting.
To build the simulator from this data, we first covered the city in a square 20 by 20 grid, and extracted Poisson parameters where is an order start tile, is an order end tile, and is the time of day in hours. This results in million parameters which we use specify an order generation process. The order generation parameters for the most active and the most quiet hours in the day averaged over all 30 days. The price of an order is determined by the trip distance.
We determine the driver generation parameters using a similar approach as we used for the order parameters. A driver is counted to have become active when it serves its first order during the day. The drivers remain active for six hours after they have been created. In the initial testing of the environment, we found that using the raw driver generation parameters results in enough drivers being created that even a randomly initialized policy is able to achieve above 99% ratio of orders served. The dataset only records orders which were served by drivers, so it seems natural that the served ratio using the unscaled parameters would be high. The approximations we make about the duration drivers remain active and the speed of the drivers could also be contributing to the high served ratio. In order to set up the environment such that there is room for improvement, we scale the driver generation parameters down to 7%. The driver generation parameters are further scaled down by a factor of 0.5, which we also apply to the order generation parameters, to lower the computational demands.
To handle the large scale of the domain with close to 100000 orders per day served by over a 1000 drivers, we use a parallelized implementation of PPO. Using parallel environments allows us to sample shorter trajectories in each environment to achieve the same update batch size. Sampling fewer consecutive transitions between each update speed up the training. We use 10 parallel environments and correspondigly reduce the steps per epoch hyperparameter to 400. We also raise the policy learning rate to and value function learning rate to .
The training curves from the experiments are presented in Figure 7. Training with either system and driver-centric approaches results in policies that clearly outperform the baselines. The best performing learned policy is the driver-centric PPO, which achieves the evaluation score of . For system-centric PPO the evaluation score is and the best performing baseline, MRM-random, scores . The driver-centric approach outperforms the system-centric by a margin of close to . It is evident from the evaluation score and the learning curves that learning with the system-centric rewards is challenging, possibly even more so as in the Historical Orders domain.
We performed a detailed empirical study of reinforcement learning approaches to multi-driver vehicle dispatching and repositioning problems. We studied driver-centric and system-centric reward definitions and trained policies using DQN and PPO algorithms. We found DQN to be more sample efficient than PPO. On the other hand PPO was better able to learn on system-centric rewards. Central to all of our approaches was the network architecture we presented, which leverages a global representation of state processed using attention mechanisms. We found that, while one can construct environments where the system-centric approach is superior, typically driver-centric is better or at least competitive with the system-centric approach. Furthermore we applied these methods to environments built using real dispatching data, and found that driver-centric approach is able to consistently beat myopic dispatching and repositioning strategies.
-  (2016) Neural combinatorial optimization with reinforcement learning. arXiv preprint arXiv:1611.09940. Cited by: §I, §II.
-  (2018) Machine learning for combinatorial optimization: a methodological tour d’horizon. arXiv preprint arXiv:1811.06128. Cited by: §II.
The dynamics of reinforcement learning in cooperative multiagent systems.
Proceedings of the AAAI conference on artificial intelligence, Vol. 1998, pp. 746–752. Cited by: §II.
-  (2002) A guide to vehicle routing heuristics. Journal of the Operational Research society 53 (5), pp. 512–522. Cited by: §I.
-  (1998) Elevator group control using multiple reinforcement learning agents. Machine learning 33 (2-3), pp. 235–262. Cited by: §II.
-  (1959) The truck dispatching problem. Management science 6 (1), pp. 80–91. Cited by: §I.
-  (2018) Learning heuristics for the tsp by policy gradient. In Integration of Constraint Programming, Artificial Intelligence, and Operations Research, W. van Hoeve (Ed.), Cham, pp. 170–181. Cited by: §II.
-  (2017) Didi launches gaia initiative to facilitate data-driven research in transportation. Note: url: http://www.didichuxing.com/en/press-news/bgcitgfc.html Cited by: §IV-D.
-  (2010-Sep.) An agent-based learning towards decentralized and coordinated traffic signal control. In 13th International IEEE Conference on Intelligent Transportation Systems, Vol. , pp. 665–670. External Links: Cited by: §II.
-  (2019) Attention, learn to solve routing problems!. In International Conference on Learning Representations (ICLR), Cited by: §II.
-  (2019) WWW ’19: the world wide web conferenceWWW ’19: the world wide web conference. Cited by: §II.
-  (2016) Asynchronous methods for deep reinforcement learning. In International conference on machine learning, pp. 1928–1937. Cited by: §III-D.
-  (2014) Recurrent models of visual attention. In Advances in neural information processing systems, pp. 2204–2212. Cited by: §II, §III-C.
-  (2015) Human-level control through deep reinforcement learning. Nature 518 (7540), pp. 529. Cited by: §III-D.
-  (2018) Reinforcement learning for solving the vehicle routing problem. In Advances in Neural Information Processing Systems 31, S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Eds.), pp. 9839–9849. External Links: Cited by: §I, §II.
-  (2018) MOVI: a model-free approach to dynamic fleet management. In IEEE INFOCOM 2018-IEEE Conference on Computer Communications, pp. 2708–2716. Cited by: §II.
-  (2007) A general heuristic for vehicle routing problems. Computers & operations research 34 (8), pp. 2403–2435. Cited by: §I.
-  (2015) High-dimensional continuous control using generalized advantage estimation. arXiv preprint arXiv:1506.02438. Cited by: §III-D, §IV.
-  (2017) Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347. Cited by: §III-D, §III-D.
-  (2002) The vehicle routing problem. SIAM. Cited by: §I.
-  (2015) Pointer networks. In Advances in Neural Information Processing Systems, pp. 2692–2700. Cited by: §II.
-  (2018) Deep reinforcement learning with knowledge transfer for online rides order dispatching. In 2018 IEEE International Conference on Data Mining (ICDM), pp. 617–626. Cited by: §II.
-  (2018) Large-scale order dispatch in on-demand ride-hailing platforms: a learning and planning approach. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 905–913. Cited by: §II.
-  (2017) A taxi order dispatch model based on combinatorial optimization. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 2151–2159. Cited by: §II, §IV.