Neural Path Planning: Fixed Time, Near-Optimal Path Generation via Oracle Imitation

04/25/2019 ∙ by Mayur J. Bency, et al. ∙ 32

Fast and efficient path generation is critical for robots operating in complex environments. This motion planning problem is often performed in a robot's actuation or configuration space, where popular pathfinding methods such as A*, RRT*, get exponentially more computationally expensive to execute as the dimensionality increases or the spaces become more cluttered and complex. On the other hand, if one were to save the entire set of paths connecting all pair of locations in the configuration space a priori, one would run out of memory very quickly. In this work, we introduce a novel way of producing fast and optimal motion plans for static environments by using a stepping neural network approach, called OracleNet. OracleNet uses Recurrent Neural Networks to determine end-to-end trajectories in an iterative manner that implicitly generates optimal motion plans with minimal loss in performance in a compact form. The algorithm is straightforward in implementation while consistently generating near-optimal paths in a single, iterative, end-to-end roll-out. In practice, OracleNet generally has fixed-time execution regardless of the configuration space complexity while outperforming popular pathfinding algorithms in complex environments and higher dimensions



There are no comments yet.


page 1

page 2

page 5

page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Being able to come up with a quick and accurate motion plan is critical to robotic systems. Motion planning involves finding a connection between two locations while avoiding obstacles and respecting bounds placed on the robot’s movement. The majority of work in solving the motion planning problem involves online computation of graphical or grid search strategies that scale poorly with dimensions, or sampling based strategies that scale better with dimensions but are highly dependent on the complexity of the environments. Furthermore, a fundamental trade-off has existed with available algorithms — the trade-off between finding the optimal solution and finding a feasible solution quickly.

Fig. 1: A composite image of the Baxter robot in action. The robot is shown executing a sequence of joint commands generated by OracleNet for picking up objects from the shelf.

The main contribution of this paper is a novel approach to the general motion planning problem that leverages a neural-network based path generator, and produces feasible paths in fixed time that mimic an oracle algorithm (one that can always generate the optimal paths across the entire configuration space of the robot/environment, for any start or end goal). Our approach leverages the Recurrent Neural Network (RNN) in order to mimic the stepwise output of an oracle planner in a predefined environment, moving from the start to the end location in a relatively smooth manner. Several important advantages of OracleNet that are demonstrated in this paper: (1) it generates extremely fast and optimal paths online; (2) it offers a valid path if one exists with probabilistic completeness; (3) it has consistent performance regardless of the configuration space complexity; and (4) it scales almost linearly with dimensions

. We demonstrate the results of our method on a point-mass robot, 3, 4, 6, and 7-degrees of freedom (DOF) robots. Because our algorithm scales close to linear with increase in dimensions, making it significantly faster and more efficient for online motion planning for non-trivial problems, compared to the polynomial or worse time complexity of popular motion planners.

Training of OracleNet is required on each new environment, making it suitable for complex robots operating in a largely static space (such as a warehouse, shopping mall, airport terminal, etc.). Small, local perturbations are allowable given retrain and repair strategies that retain probabilistic optimality and completeness guarantees. Although not investigated in this work, OracleNet can also be made to adapt to dynamic environments by contextualizing path generation on obstacle representation in addition to the start and goal. This is considered in [1], where we expand on the original concept of neural motion planning discussed in this paper to generate paths efficiently in a variety of unseen environments and compares against the state-of-art planners. However, encoding environment information, especially for multi-DOF systems, is far from trivial [1]

because 1) the robot’s surrounding environment and joint-space constitute entirely different distributions and therefore mapping from one to the other is challenging and a research problem of its own; 2) the environment information is usually taken as raw point-cloud data, encoding which into an interpretable feature space is hard as the representation needs to respect the permutation invariance of the input point cloud. OracleNet implicitly learns an obstacle encoding from its training set of expert trajectories, thereby by-passing expensive environment encoding. An active learning strategy may also be employed that will enable OracleNet to be retrained to incorporate changes in the environment in just a few iterations. In general, this approach offers the starting concept of formulating the path planning problem with a sequential neural network-based solver (i.e.

neural motion planning) in which many algorithmic variants can be considered, including those that extend to unseen environments and dynamic environments.

Ii Related Work

A range of techniques to solve the motion planning problem has been proposed in the past two decades, from algorithms that emphasize optimality over computational expense, to those that trade-off computational speed with optimality [2, 3, 4]. Traditional algorithms such as A* [5] that search on a connected graph or grid, while fast and optimal on small grids, take exponentially longer to compute online with increasing grid sizes and environment complexity. Since the results presented in this work focus on static environments, precomputed roadmap-based planners can also be considered. Probabilistic Roadmaps (PRM), while appearing to solve the sample-size issue by decoupling the number of samples required to construct a graph from the dimensionality of the space, suffer from the same fate of requiring an exponential number of points in dimension to maintain a consistent quality. In fact, any sampling scheme, random or deterministic, has been shown to suffer from this [6]. Preliminary results for benchmarking PRM have been provided in 1. Sampling-based strategies such as RRT have better computational efficiency for searching in high dimensional spaces [7] but get slowed down by their “long tail” in computation time distribution in complex environments. Apart from using grid-based and sampling-based motion planners, optimizing over trajectories has also been proposed, with approaches such as using potential fields to guide a particle’s trajectory away from obstacles [8] and reformulating highly non-convex optimization problems to respect hard constraints [9]. A review of recent algorithms and performance capabilities of motion planners can be found in [10].

The challenge of creating and optimizing motion plans that incorporate the use of neural networks has long been a problem of interest, though computational efficiency in solving for deep neural networks has only recently made this a practical avenue of research. An early attempt aimed to link neural networks to path planning by specifying obstacles into topologically ordered neural maps and using neural activity gradient to trace the shortest path, with neural activity evolving towards a state corresponding to a minimum of a Lyapunov function [11]

. More recently, a method was developed that enables the representation of high dimensional humanoid movements in the low-dimensional latent space of a time-dependent variational autoencoder framework


Reinforcement Learning (RL) approaches have also been proposed for motion planning applications [13, 14]. Recently, a fully differentiable approximation of the value-iteration algorithm was introduced that is capable of predicting outcomes that involve planning-based reasoning [15]

. However, their use of Convolutional Neural Networks to represent this approximation limits their motion planning to only 2D grids, while generalized motion planning algorithms can be extended to arbitrary dimensions. RL assumes that the problem has the structure of a Markov Decision Process where the agent attempts to solve the problem through a trial-and-error based interaction with the real environment. On the other hand, classical motion planning algorithms take in a full state information map as a part of the planning problem and output a solution without a single interaction with the real environment. The algorithm presented in this work leverages the assumptions used in the latter and thus is different from RL-based approaches.

Learning to generate motion plans has also been considered via a Learning-from-Demonstration (LfD) approach. Using an “expert” (usually human) to provide demonstrations of desired trajectories, LfD methods are able to generalize within the set of demonstrations an approximate, underlying sequence or policy that reproduces the demonstrated behavior. LfD has been successfully applied in various situations that involve challenging dynamical systems or nuanced activities [16, 17]

. Our path generating algorithm can be considered an extension of LfD since the lines between motion planning, imitation learning, and model predictive control are getting blurred with advancements in machine intelligence.

Fig. 2:

An “unfolded through time” strategy for motion planning is proposed in OracleNet. Note that within this RNN structure, the Long Short Term Memory (LSTM) hidden layer weights are shared across timesteps as the inputs get iteratively updated and concatenated.

Iii Methods

Iii-a Problem Definition

In this work, we use the standard definition of configuration spaces (c-spaces) to construct the environment in which our motion planning algorithm operates. For a robot with degrees-of-freedom (DOF), the configuration space represents each DOF as a dimension in its coordinate system. Each -dimensional point in the c-space represents the joint angles of the robot and therefore, the full configuration of the robot in the real world. Due to this property, motion planning in c-spaces is simpler than in geometric spaces. A motion planning task involves connecting two points in a -dimensional c-space, with obstacles mapping from Cartesian space to c-space in a nonlinear fashion through forward-kinematic collision checks [18]. A necessary assumption for our algorithm is knowledge of the configuration space, which is defined by a given robot present in a given static constrained environment. For example, these problems arise in solving for optimal routes in a large network of roads or navigating through the lanes of a crowded warehouse.

Let be the c-space. Let be the obstacle region, such that is an open set, and denote the obstacle-free space as , where denotes the closure of a set. The initial start point and the goal , both elements of , are provided as query points. The objective here is to find a collision-free path between and in the c-space. Let be defined as a discrete sequence of waypoints. is considered valid if it is continuous, collision free (each waypoint in the generated path should lie in ), and feasible for some finite t).

Iii-B Proposed Algorithm

We propose to solve the problem of generating goal-oriented path sequences by passing in a goal location as an auxiliary input at each step of the prediction, which provides a “reminder” to the network about where it should ultimately converge. At each step, the input vector is concatenated with the desired goal location and the resulting augmented input vector is used to train an RNN model. The RNN is trained on optimal trajectories that span the entire configuration space, which allows it to internalize an oracle-mimicking behavior that generates optimal path sequences during rollouts. The network comprises of stacked Long Short Term Memory (LSTM) layers that preserve information over a horizon of outputs

[19], with the output layer being fully connected to the final LSTM hidden layer. In this work we concern ourselves with fixed short-term memory; our experiments showed that using LSTMs resulted in better performance over vanilla RNNs. Mean Squared Error between the predicted output and the teaching signal is used to train the network. The number of layers was decided on by empirically converging to an appropriate size based on the dimensionality of the problem to be solved. Increasing the number of layers as we increase the dimension of the c-space to capture additional degrees of freedom in the training set lead to better performance. For any finite time trajectory of a given -dimensional dynamical system, a practical limit in expanding network size can be approximately realized by the internal state of the output units of a continuous time recurrent neural network with output units, some hidden units, and an appropriate initial condition [20] . Exact numbers for network size and depth are provided in the Section IV.

Iii-C Training Set Creation and Offline Training

A training set consisting of a number of valid paths created by an “expert” planner, which we call the Oracle. We used the A* to generate an optimal set of paths for training. The c-space is sampled to create a graph with connected nodes. Two nodes are randomly selected without replacement (based on a uniform distribution) from the set of nodes present in

and A* is executed to find the optimal path connecting them. This process is repeated times to obtain a training set consisting of “expert” paths. Each generated path is split into their composing waypoints to make each individual waypoint represent a sample in the training set. The teaching signal corresponding to each sample then becomes the next waypoint in the path sequence. That is, if is a path with waypoints, the path is split into {, } and the corresponding teaching signals become {, }. If we assume that is the goal point in a sequence, then the auxiliary input is concatenated as , where ranges from 0 to . The LSTM network is trained on the .

Iii-D Online Execution through Bi-directional Path Generation

Given a trained network, for testing, we select two points and (not from the training set) from and attempt to roll out a path connecting them while avoiding obstacles. The network generates a sequence of waypoints until a final connection to occurs to complete the path. After the process is terminated, all the sequentially generated outputs are formatted as waypoints in the generated path. To make the path generating process more robust, bi-directional path generation is used. We start the generation process from and points simultaneously and make the two branches grow towards each other. The process is terminated when the two branches meet, and the branches are then stitched together to form a complete path. Fig. 3 (b) shows the bidirectional stepping behavior. Instead of forcing the path to “grow” towards an arbitrarily selected fixed goal point, the network now has the option to target a constantly shifting goal point (the current point of the other branch). This increases the chance of the convergence point falling with the Oracle paths on which OracleNet was trained, thus increases feasibility and success rates, while having no impact on path roll-out time.

Fig. 3: (a) An example path generated by OracleNet, A*, and RRT*. (b) Example of bi-directional path generation (not rewired) between start (red) and goal (green) configurations. (c) Example of bi-directional, repaired, and rewired path generation. Figure inset shows the repair module in action (magenta line). After the path is repaired and converges successfully, the rewire module is called to remove superfluous nodes and any kinks the repairing may have introduced (black line).

Iii-E Repair and Rewire

It is to be expected that OracleNet will never yield an exact duplication of an Oracle’s behavior. Also, since the expert demonstration paths are sampled randomly on the c-space, there may be regions in the c-space that the network has never seen in the dataset. Because of these unseen regions, the generated waypoints may not adhere strictly to obstacle boundaries and cut corners through obstacle regions at times. This in practice happens less than 3% in all paths generated by the network. Naive methods may exist such as obstacle padding. However, we propose a

repair module to fix violating waypoints as they appear online while generating the path. The repairing strategy used is straightforward to implement. When current waypoint is generated inside an obstacle region, a direction is randomly selected at a step distance from x(t-1) reach a new candidate . Random samples are taken until the first feasible is found. then replaces and then OracleNet continues. Since the random search is initiated from which, by definition, is a valid point, the chances there will be at least one from which OracleNet can proceed are very high. With an appropriately chosen and testing the connectivity of two points with a collision checker (analogous to the “local planner” in PRM), feasible paths are generated in probabilistic completeness.

The repair methods above, along with general network noise, will result in paths that can be non-smooth. To deal with this, we propose a rewiring process that removes unnecessary nodes in the paths by evaluating if a straight trajectory connecting two non-consecutive nodes in the path is collision-free. Similar to the rewiring algorithm used in RRT*, a lightweight implementation of this algorithm has very little processing overhead and thus can be used without a noticeable increase in path generation times. It is interesting to note that the final path thus generated and processed has a practical time complexity of (as shown through results in Section IV). To maintain fairness and consistency in optimality comparisons, the proposed rewiring technique is also applied to the output of A* searches in all our experiments. The full procedure for OracleNet in online path generation is outlined in Algorithm 1.

1 Trained OracleNet while  do
2       if  then
Algorithm 1 OracleNet (Online Path Rollout)

Iv Results

Fig. 4: 8 environments are used in the 2D Gridworld experiments. The top row has “simple” environments, numbering from the left, while the bottom row has “difficult” environments. Each environment shows a single roll-out of OracleNet. Unlike pathfinding algorithms, no expanding search is required.
Completion Time Optimality (Ratio of Path Lengths)
Environment A* (s) RRT* (s) OracleNet (s) OracleNet / A* OracleNet / RRT*
Simple 1 0.08 (0.06) 1.98 (3.68) 0.13 (0.18) 0.97 (0.08) 0.84 (0.16)
Simple 2 0.09 (0.07) 1.23 (1.76) 0.24 (0.18) 0.96 (0.03) 0.86 (0.11)
Simple 3 0.07 (0.06) 0.93 (2.9) 0.16 (0.19) 0.96 (0.04) 0.86 (0.10)
Simple 4 0.08 (0.05) 1.53 (2.43) 0.18 (0.20) 0.98 (0.05) 0.87 (0.12)
Difficult 1 0.07 (0.05) 2.69 (3.54) 0.18 (0.10) 0.97 (0.07) 0.87 (0.10)
Difficult 2 0.07 (0.05) 3.67 (6.31) 0.18 (0.12) 0.99 (0.05) 0.88 (0.12)
Difficult 3 0.09 (0.06) 5.17 (11.57) 0.17 (0.12) 0.96 (0.22) 0.87 (0.12)
Difficult 4 0.05 (0.04) 8.79 (12.81) 0.18 (0.09) 0.96 (0.17) 0.89 (0.11)

Values are listed as “mean (standard deviation)” for 8 different environments.

Lower is better. Below 1 means that OracleNet produces shorter paths.

TABLE I: Speed and Optimality of OracleNet benchmarked against A* and RRT* in a 2D environment

To appropriately evaluate the performance and capability of OracleNet, we test it on a number of distinct environments. The experiments were conducted in a 2D Gridworld with a point robot having translation capabilities only, 3-link, 4-link, and 6-link robot manipulators. For all experiments presented here, training is accomplished with Tensorflow and Keras, a high-level neural network Python library

[21], with a single NVIDIA Titan Xp used for GPU acceleration.

Iv-a 2D Gridworld

Fig. 4 shows snapshots of 8 environment examples, 4 considered “simple” environments for popular motion planning strategies such as RRT*, and 4 “difficult” environments. An environment is considered simple if the obstacles are convex and widely spaced apart, while difficult environments consist of either a large number of obstacles or highly non-convex obstacles forming narrow passageways. Each continuous space environment is 100 units in length and width. A* is run on a unit grid of 100 x 100 on this environment to generate 20,000 valid Oracle paths for training OracleNet. This set was split in accordance to the 80-20 rule, with 20 being kept for testing to control overfitting. The network architecture consists of 4 LSTM layers each of 256 hidden units.

Fig. 5: Example plots of successful paths generated on 3-link, 4-link, and 6-link robots (from the left) using our algorithm, showing the trail of the links as it makes its way from its initial joint configuration (marked as green).

To benchmark our algorithm’s performance with existing motion planning algorithms, we use RRT* and A*. While these planners are not among the most efficient (and we do compare neural planning against the state of art in [1]), they are both widely used and accessible, and demonstrate the advantages of training a neural planner over using the planning algorithms during runtime. Three performance metrics are used: success rate, roll-out time, and path optimality. A generated path is considered successful if none of the waypoints encroach into obstacle region. Roll-out time measures the time taken for the network to generate waypoints from start to goal (or, in the bidirectional case, the time taken for both branches to meet). Path optimality is simply the fraction of the path length generated by OracleNet when benchmarked against paths generated by A* and RRT* respectively. 1000 randomly initialized trials were conducted for each of the 8 environments. Table I shows that OracleNet manages to be comparable to A* and faster than RRT* even for a small 2D grid while being slightly more optimal in both cases. This is due to the rewiring module and the network being able to generate points in continuous space as opposed to being restricted to a discrete grid in A*. Thus, the proposed algorithm may be a suitable alternative even for small grids with limited connectivity, if path optimality is the priority.

Fig. 6: A histogram of the test cases used to evaluate performance on a 3-link arm. Refer Table II

for means and standard deviations. Note the Gaussian shape of the distribution for OracleNet, compared to the left-skewed exponential distributions of A* and RRT*. Higher standard deviations for A* and RRT* cause the generation times to have a much wider spread, while OracleNet’s much tighter spread indicates its consistency in near fixed-time execution.

Completion Time Optimality (Ratio of Path Lengths)
Environment A* (s) RRT* (s) OracleNet (s) OracleNet / A* OracleNet / RRT*
3-link (3D) 2.707 (4.28) 3.83 (6.68) 0.22 (0.21) 1.02 (0.12) 0.75 (0.20)
4-link (4D) 61.87 (95.07) 18.21 (14.76) 1.18 (0.87) 0.99 (0.14) 0.86 (0.11)
6-link (6D) 727.56 (105.11) 29.32 (6.25) 1.24 (0.72) 0.95 (0.15) 0.85 (0.17)

Values are listed as “mean (standard deviation)”.

Below 1 means that OracleNet produces shorter paths.

TABLE II: Performance of RNN-based motion planners in higher dimensions benchmarked against A* and RRT*

Iv-B Multi-Link Planar Manipulators

To demonstrate the extensibility of the proposed method to higher dimensions, OracleNet was tested on 3-link, 4-link, and 6-link manipulators. The base link has movement range 0 to 2 while the subsequent links can move between to . For the 3-link manipulator, we discretized the 3-dimensional joint angle c-space into a 3D uniform grid with 50 nodes on each axis, resulting in a total of = 125,000 uniformly spaced nodes. For the 4-link and 6-link cases, the corresponding dimensional c-space is discretized into 40 and 10 uniformly spaced nodes per axis, respectively. As in the 2D case, A* is used to generate the training set. To get an accurate representation of the complete c-space in the training set (which directly correlates to the increased number of nodes in the grid used here for training), we use 400,000 for the 3-link case, and 1 million each for the 4-link and 6-link cases. Keeping in mind the increased number of dimensions to learn, we updated the architecture to have 6 layers with 256 units each for the 3-link case, and 6 layers with 400 units each for both 4-link and 6-link cases. For our performance evaluation, we randomly generate 1000 pairs of start and goal locations in continuous c-space. Paired with the repair and rewire modules discussed in the previous sections, we observed a 100 success rate in finding feasible paths. Generation times and path optimality is benchmarked against A* and RRT*. Table II shows that OracleNet scales much better than A* and RRT*. An interesting observation to note is the low standard deviation of path generation times for OracleNet. This is further expanded on in the histogram of the test cases shown in Fig. 6. This is indicative of the consistency of OracleNet in producing its paths across the entire c-space, whereas A* and RRT* are heavily influenced by the relative locations of the query points and the obstacles. More about this is discussed in Section V.

Iv-C 7-DOF Dual-Arm Baxter Robot

As a final demonstration of OracleNet’s capabilities, we show its application on a humanoid dual-arm Baxter robot.We train the network to output a sequence of 7 DoF joint-angles (for one arm) for the robot to follow from a starting configuration to a goal configuration. The robot, after training, is expected to generate a sequence of feasible configurations (if they exist) connecting given start and goal configurations. Due to the high dimensionality of the configuration space, it becomes infeasible to discretize it to a grid of reasonable resolution and use graph search algorithms such as A* to generate optimal paths. Instead, we rely on Open Motion Planning Library (OMPL) to generate a set of paths between randomized valid configurations using RRT-Connect [22]

. RRT-Connect is an efficient sampling-based planner that combines RRT’s sampling scheme with a simple greedy heuristic to generate quick single-query paths. Fig.

7 shows a simulated version of the robot performing a pick-place-task. OracleNet generated the sequence of joint angles necessary for the robot arms to move towards the given object locations. Fig. 1 shows this being achieved with the physical Baxter robot as well. A relatively small dataset of 40,000 paths is generated and train on it using a model with a reduced number of layers (2 and 3 layers with 400 units each were experimented with). Due to the significantly constrained work-space in which the Baxter has to operate, connecting arbitrary configurations for testing may not always lead to viable paths. Even RRT-Connect, the “expert planner” used here, frequently failed to generate viable paths before timing-out. Despite the lower-quality dataset available for this experiment, utilizing its generalizing ability and the fail-safe measures described previously, OracleNet was able to generate smooth paths connecting the chosen configurations with mean and standard deviation of generation times as 0.56 and 0.39 seconds respectively.

Fig. 7: A composite image of the Baxter executing a pick-and-place task with manually controlled grasp. The trail of the arms show the path taken by them to reach their respective objects. The right arms plans towards the duck while the left arm plans towards the green cuboid.

Iv-D Scalability and Effects of Path Length

Fig. 9 shows the scalability of OracleNet with dimensions. In conjunction with Table I, it shows that OracleNet provides minimal improvements to readily accessible algorithms like A* in low dimensional cases with small grid sizes (low resolution). In spaces with dimensions greater than 2, OracleNet scales far better, as A* (paired with a Euclidean distance heuristic) scales exponentially worse with dimensions, and RRT* (depending on the trade-off selected between step size and path optimality, scales polynomial at best). Results presented in Table II and Fig. 9 present the trends that show a modest increase in execution time for OracleNet while path optimality rivals A* and far outstrips RRT*.

Fig. 8 shows the spread of generation times of paths when arranged as a function of path length (demonstrated on the 3-link arm). OracleNet has a linear relationship to how far away the query points are located, meaning that stepping through the environment is done consistently with a fixed time. For pathfinding algorithms such as A*, the farther apart the query points are, the more chances are a number of nodes required to be explored increases exponentially (polynomial with an appropriately chosen admissible heuristic). The fixed-time behavior of OracleNet is especially valuable in time-critical scenarios where a feasible path must be known in fixed time.

Fig. 8: OracleNet and A* path generation times plotted as a function of path lengths for the 3-link robot presented here. Note the near fixed-time generation times for OracleNet, irrespective of where the query points are located in the entire c-space.

V Discussion

The three main qualities required of any successful motion planner are feasibility, speed, and optimality. While feasibility is usually non-negotiable, speed and optimality can often be traded for each other depending on what the priorities are. This work attempts to achieve both high speed as well as optimality while simultaneously preserving feasibility. The experiments performed in the various environments presented cover a range of cases that support this effort.

Using a discrete uniform grid to build the training set allows us to estimate the efficiency of the dataset and the generalizing ability of the network. The fact that it is possible to successfully generate paths between two continuous free floating points when it is trained only on a sparse discrete grid indicates generalizing ability of the network. The 6-link experiment is a good example of this. Even with just 10 uniformly spaced discrete samples per axis, the network managed to find optimal collision-free paths connecting continuous points sampled arbitrarily in the c-space.

Another effect of network size beyond the required training data directly correlate with path generation times, an increasing network size will increase the path generation time at a comparable rate. While overfitting is avoided (as described in Section III), larger networks produce minimal performance improvements, and it is worth empirically honing it to an appropriate network size. A complementary strategy to potentially reduce network size while retaining performance and potentially improving robustness is using dropout [23].

A unique property of OracleNet is with regards to real-time execution. Unlike any other motion planning strategy (with the exception of the potential field strategies), a movement can be initiated immediately before a path through the environment is found. Because OracleNet encodes optimal behaviors, one can execute OracleNet under the belief that any step it takes will move towards the right direction. Issues such as repair and rewiring can be resolved by considering an N-step horizon, which is a typical strategy in online motion planning and specifically re-planning. Thus, in practice, one can expect that a robot can move and react instantaneously once given its goal state.

The speed and invariance of performance to environments do come at a cost, namely the creation of a dataset, and time and computation power to accurately generate paths. Taking advantage of multi-processor parallelization, for the training set sizes posted in Section IV, the dataset creation times ranged from around 2 hours for the 2D grids to around 12 hours for the 6D cases. For the training times, with GPU acceleration and for the network sizes mentioned in Section IV, the training times ranged from around 4 hours till convergence for the smaller 2D cases to 10 hours for the larger 6D cases. This cost is easily accepted if the algorithm is viewed as an extremely fast way to create online optimal and feasible motion plans for any start and goal state in a static environment, given that the process of creating training data and training of the network occurs offline. As mentioned previously, potential real-world applications of this algorithm include warehouse scenarios, robot picking and shelving, and custodial robots.

Fig. 9: A comparison of average path generation times for our algorithm, A*, and RRT* for all the cases presented here. Because of the unreasonably high complexity of running A* in a 6-dimensional c-space with full connectivity, it was chosen not to be included. The scalability for online path generation is far better for OracleNet and produces near-optimal paths through the entire space which cannot be done practically with other competing methods.

Given a foundational framework for mimicking Oracles using neural networks, several extensions can be pursued. Adapting to dynamic environments, unseen environments via transfer learning, improving sampling strategies for Oracles (i.e. training data selection) are considered in

[1], and can be extended towards a fast and efficient local planner component in task-based planners [24].


  • [1] A. H. Qureshi, A. Simeonov, M. J. Bency, and M. C. Yip, “Motion planning networks,” arXiv preprint arXiv:1806.05767, 2018.
  • [2] M. Rickert, O. Brock, and A. Knoll, “Balancing exploration and exploitation in motion planning,” in Robotics and Automation, 2008. ICRA 2008. IEEE International Conference on.   IEEE, 2008, pp. 2812–2817.
  • [3] A. H. Qureshi and Y. Ayaz, “Potential functions based sampling heuristic for optimal path planning,” Autonomous Robots, vol. 40, no. 6, pp. 1079–1093, 2016.
  • [4] ——, “Intelligent bidirectional rapidly-exploring random trees for optimal motion planning in complex cluttered environments,” Robotics and Autonomous Systems, vol. 68, pp. 1–11, 2015.
  • [5] P. E. Hart, N. J. Nilsson, and B. Raphael, “A formal basis for the heuristic determination of minimum cost paths,” IEEE transactions on Systems Science and Cybernetics, vol. 4, no. 2, pp. 100–107, 1968.
  • [6] S. M. LaValle, M. S. Branicky, and S. R. Lindemann, “On the relationship between classical grid search and probabilistic roadmaps,” The International Journal of Robotics Research, vol. 23, no. 7-8, pp. 673–692, 2004.
  • [7] S. M. LaValle, “Rapidly-exploring random trees: A new tool for path planning,” 1998.
  • [8] O. Khatib, “Real-time obstacle avoidance for manipulators and mobile robots,” in Autonomous robot vehicles.   Springer, 1986, pp. 396–404.
  • [9] N. Ratliff, M. Zucker, J. A. Bagnell, and S. Srinivasa, “Chomp: Gradient optimization techniques for efficient motion planning,” in Robotics and Automation, 2009. ICRA’09. IEEE International Conference on.   IEEE, 2009, pp. 489–494.
  • [10] D. González, J. Pérez, V. Milanés, and F. Nashashibi, “A review of motion planning techniques for automated vehicles,” IEEE Transactions on Intelligent Transportation Systems, vol. 17, no. 4, pp. 1135–1145, 2016.
  • [11] R. Glasius, A. Komoda, and S. C. Gielen, “Neural network dynamics for path planning and obstacle avoidance,” Neural Networks, vol. 8, no. 1, pp. 125–133, 1995.
  • [12] N. Chen, M. Karl, and P. van der Smagt, “Dynamic movement primitives in latent space of time-dependent variational autoencoders,” in Humanoid Robots (Humanoids), 2016 IEEE-RAS 16th International Conference on.   IEEE, 2016, pp. 629–636.
  • [13] S. Levine and V. Koltun, “Guided policy search,” in

    International Conference on Machine Learning

    , 2013, pp. 1–9.
  • [14] D. Silver, T. Hubert, J. Schrittwieser, I. Antonoglou, M. Lai, A. Guez, M. Lanctot, L. Sifre, D. Kumaran, T. Graepel et al., “Mastering chess and shogi by self-play with a general reinforcement learning algorithm,” arXiv preprint arXiv:1712.01815, 2017.
  • [15] A. Tamar, Y. Wu, G. Thomas, S. Levine, and P. Abbeel, “Value iteration networks,” in Advances in Neural Information Processing Systems, 2016, pp. 2154–2162.
  • [16] P. Abbeel, A. Coates, and A. Y. Ng, “Autonomous helicopter aerobatics through apprenticeship learning,” The International Journal of Robotics Research, vol. 29, no. 13, pp. 1608–1639, 2010.
  • [17] S. Calinon, F. D’halluin, E. L. Sauser, D. G. Caldwell, and A. G. Billard, “Learning and reproduction of gestures by imitation,” IEEE Robotics & Automation Magazine, vol. 17, no. 2, pp. 44–54, 2010.
  • [18] N. Das, N. Gupta, and M. Yip, “Fastron: An online learning-based model and active learning strategy for proxy collision detection,” in Conference on Robot Learning, 2017, pp. 496–504.
  • [19] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation, vol. 9, no. 8, pp. 1735–1780, 1997.
  • [20] K.-i. Funahashi and Y. Nakamura, “Approximation of dynamical systems by continuous time recurrent neural networks,” Neural networks, vol. 6, no. 6, pp. 801–806, 1993.
  • [21] F. Chollet et al., “Keras,”, 2015.
  • [22] J. J. Kuffner and S. M. LaValle, “Rrt-connect: An efficient approach to single-query path planning,” in Robotics and Automation, 2000. Proceedings. ICRA’00. IEEE International Conference on, vol. 2.   IEEE, 2000, pp. 995–1001.
  • [23] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” The Journal of Machine Learning Research, vol. 15, no. 1, pp. 1929–1958, 2014.
  • [24] A. Krontiris and K. E. Bekris, “Dealing with difficult instances of object rearrangement.” in Robotics: Science and Systems, 2015.