Virtual-to-Real-World Transfer Learning for Robots on Wilderness Trails

01/17/2019 ∙ by Michael L. Iuzzolino, et al. ∙ University of Colorado Boulder 14

Robots hold promise in many scenarios involving outdoor use, such as search-and-rescue, wildlife management, and collecting data to improve environment, climate, and weather forecasting. However, autonomous navigation of outdoor trails remains a challenging problem. Recent work has sought to address this issue using deep learning. Although this approach has achieved state-of-the-art results, the deep learning paradigm may be limited due to a reliance on large amounts of annotated training data. Collecting and curating training datasets may not be feasible or practical in many situations, especially as trail conditions may change due to seasonal weather variations, storms, and natural erosion. In this paper, we explore an approach to address this issue through virtual-to-real-world transfer learning using a variety of deep learning models trained to classify the direction of a trail in an image. Our approach utilizes synthetic data gathered from virtual environments for model training, bypassing the need to collect a large amount of real images of the outdoors. We validate our approach in three main ways. First, we demonstrate that our models achieve classification accuracies upwards of 95 our synthetic data set. Next, we utilize our classification models in the control system of a simulated robot to demonstrate feasibility. Finally, we evaluate our models on real-world trail data and demonstrate the potential of virtual-to-real-world transfer learning.



There are no comments yet.


page 1

page 2

page 3

page 4

page 5

page 6

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Robots have shown significant aptitude in data-gathering and inspection tasks across a variety of domains. Aerial robots are especially adept at such tasks (see [1] for a survey). Due to substantial cognitive demands, many aerial robot systems are manned by teams of two humans, one acting as pilot and the other as mission specialist orchestrating the collection and analysis of data [2]. Extending autonomous navigation capabilities could positively impact robot operations by attenuating cognitive demands and allowing human operators to focus on other mission-critical tasks that involve high-level decision making.

Autonomous navigation of outdoor trails presents a complex, non-trivial perception and planning problem. Unlike well-defined environments, such as roadways and sidewalks in urban areas, wilderness trails consist of drastically varying features (e.g., gravel path, game trail, backcountry dirt road), traverse highly variable terrains, and span vastly differing biomes (e.g., forests, meadows, mountains), all under various seasonal and lighting conditions. Dense vegitation and large obsticles may pose significant visibility constraints, while GPS measurements may be unreliable or even unavailable [3]

. Consequently, autonomous navigation of unknown terrain and environments is an active area of research within the fields of machine learning and robotics.

Deep learning approaches are establishing state-of-the-art results for robot perception, planning, navigation tasks. However, such approaches require large, labeled training datasets that often require exhaustive human labor for collection and labeling. In many instances, collecting and labeling these datasets poses significant challenges, some of which may be insurmountable due to logistical issues. For example, search-and-rescue is particularly critical during harsh weather conditions, but it is these hazardous conditions in which it is most difficult to collect training data for data-driven approaches, such as deep learning.

In this paper, we demonstrate a deep learning approach that may mitigate these issues through the utilization of transfer learning between virtual and real-world domains. We propose a solution for training neural networks on synthetic images of virtual outdoor trails, where a neural network learns to identify the direction of the trail within an image, and demonstrate that the features learned on the virtual dataset are capable of transferring to real-world domains for trail perception. Our method alleviates the need for exhaustive real-world data collection and laborious data labeling efforts.

Ii Related Work

Our approach draws from the fields of robotic perception, computer vision, and deep learning. Below, we discuss image classification and object detection for trail perception. We then review advances in transfer learning between real-world datasets and discuss extensions of transfer learning to virtual and real-world datasets for use in robot perception and navigation tasks.

Previous efforts to solve the problem of autonomous pathfinding and navigation focused on trail segmentation using low-level features such as image saliency or appearance contrast [2, 4]. However, more recent approaches have leveraged deep learning to produce cutting-edge results for elements of robotic navigation, such as trail perception and object detection. In the work of Giusti et al. [4], a hiker was equipped with three head-mounted GoPro cameras with left-, center-, and right-facing orientations and traversed alpine regions of Switzerland for 8 hours, resulting in a dataset of 24,747 natural trail images. The camera setup allowed for automatically labeled data: images collected by the left-facing GoPro camera were labeled as left, and so on.

This dataset was used to train a convolutional neural network that learned to discriminate on salient features that best predict the most likely classification of the image. This method achieved classification accuracies of 85.2% and outperformed conventional computer vision methods, such as hue-based saliency mapping for RBF kernel SVM classification (52.3%), and is comparable the performance of humans (86.5%)

[4]. The network was qualitatively evaluated as a control system for a real-world aerial drone with a monocular camera and demonstrated moderate success.

While Giusti et al. [4] demonstrated promising results, their approach relies on real-world data collection and may thus be limited due to issues arising from battery-life, human fatigue, data collection errors due to incorrect head orientation and mislabeling of data, or seasonal availability and safety. In addition, this approach may not extend to inaccessible, novel, and/or dangerous environments, such as rugged winter trails or extraterrestrial environments (e.g., for use in robotic space exploration on Mars or the Lunar surface). A possible solution to these challenges is transfer learning, an active area of research within the deep learning community, where knowledge representations are learned in one domain and utilized to accelerate learning in a related domain. For instance, research has revealed that convolutional neural networks trained on natural images learn generalizable features, such as Gabor filters and color blobs [5]

, that form the basis of many datasets, such ImageNet

[6] datasets.

Our approach is inspired by transfer learning; however, instead of transferring from one real-world domain to another, we are interested in the notion of transferring knowledge learned in virtual environments to the real world. For example, prior work has developed a mapless motion planner for real environments by training a deep reinforcement model in synthetic settings [7]. After training in a well-defined simulation, the system converges upon an optimal set of navigational policies that are then transferred to a real-world robot capable of navigating a room with static obstacles. This work highlights the potential of virtual-to-real transfer learning in domains where a well-defined simulation is available. However, this work does not address the challenges of perception and navigation in complex environments where simulations may be lacking or non-existent. Our work in this paper further explores the potential of virtual-to-real-world transfer learning to address the challenges raised by complex domains, such as wilderness trails.

Fig. 1: Birds-eye view subsection of trails (dotted-red line) traveled by virtual camera and robot.

Iii Approach

To explore the concept of virtual and real-world trail navigation, we created a virtual environment for synthetic data collection. Below, we discuss the details of the virtual environment. Then we outline our methods for data collection, processing, and constructing three different archetypal neural networks. Finally, the last two sections describe the integration of the trained neural network with the Unity environment—a cross-platform 3D game engine—and the method used for evaluating our models on real-world data to demonstrate virtual-to-real-world transfer learning.

Iii-a Virtual Environment

To create our virtual environment, we used Unity, a game and animation engine for developing virtual interactive 3D environments. Using the built-in terrain editor and readily available 3D models of natural objects (trees, rocks, grass, etc.) from the Unity Asset Store, we assembled a virtual scene of an alpine mountain with a web of dirt trails spanning the landscape (Figure 1). The paths in the environment held many similarities to real-world trails: they branched, curved around rocky corners and wooded areas, changed elevation, and contained ambiguous trail sections.

Iii-B Data Collection

A single path in the Unity environment was randomly chosen and utilized for all data collection. A virtual robot was placed onto the path and a C# control script was attached to the robot that enabled it to deterministically traverse the path multiple times. Three cameras, each with pixel resolution, 30 frames per second (FPS), and an 80 field-of-view (FOV) were attached to the robot. The combined FOV capture for all three cameras spanned 180 with each periphery camera having 30 of overlap with the central camera (Figure 2). The camera configurations were determined in conjunction with the data capture setup in [4], GoPro camera design specifications, and the results from preliminary virtual-world data capture trials. The virtual robot’s roll and pitch were constrained to 0, with the yaw always set to a value that directed the robot toward the center of the path. The robot traversed the path and collected a total of 20,269 images (center: 6821, left: 6829, right: 6619). The screen-shot bundles were labeled as either center, left, or right—depending on which camera the images were captured from—and stored locally.

Fig. 2: Top-down view of camera configuration for data capture in Unity environment.

Iii-C Data Processing

Standard image processing practices, such as resizing and normalization, were followed. The images were resized to

pixels. This allowed for faster processing and lower memory consumption, which is especially problematic for large neural network models. We then normalized the images to account for the highly variable range of pixel values. Non-normalized data is problematic during back-propagation for most machine learning algorithms, where weight changes are computed by the accumulation of the gradient, multiplied by a scalar learning rate. With non-normalized feature vectors, the result is typically an oscillatory behavior of the gradients, as the weights of some features are over-corrected whereas others are under-corrected. Consequently, we normalized the pixel-space of our images to values between

. To normalization across a large, high-dimensional dataset, we opted to normalize on each image and per color channel, rather than across the distribution of all images in the dataset.

Training, Validation, and Test Sets: The virtual data collected via Unity was split into three sets: training, validation, and testing. The real-world dataset from [4] was utilized as an additional test set to demonstrate the transferability of features between virtual and real-world domains. The splits and distributions are presented in Table LABEL:table:data_dist.

Iii-D Model Architectures

We explored three different model architectures: standard feed-forward deep neural networks (DNN’s), convolutional neural networks (CNN’s), and recurrent neural networks (RNN’s). In the following subsections, we outline the models’ hyperparameters, and input/output structures.

Iii-D1 Deep Neural Network

The feed-forward network is outlined in Figure 3 (A). The pixel input images are flattened to a dimensional input vector and fed to the input layer of the DNN, which contains

input neurons. This architecture implemented three hidden layers (not shown in Figure


) and utilized rectified linear unit (ReLU) activation functions. The output of the last hidden last layer is sent to a final output layer that consists of three neurons, where each maps to a corresponding classification prediction of left, center, or right. A


activation is applied to the outputs, establishing a proper probability distribution over which the

argmax yields the classification prediction.

Iii-D2 Convolutional Neural Network

The architecture of this model replicates that of the CNN utilized in [4] (Figure 3, C). The pixel input images are fed into the first convolution layer, which contains 32 filters,

kernels, and a stride of 1. The convolution layer is activated by a sigmoid function and then fed to a max pool layer with kernel sizes of

and strides of 2. This block of convolution, activation, and max-pooling is repeated with each unit containing identical parameters a total of four times. The 4th max pooling layer is flattened and fed to a fully connected layer with 200 neurons, and the sigmoid activated output is fed to the final output layer containing three neurons. The output layer is identical to the DNN.

Iii-D3 Recurrent Neural Network

This architecture is depicted in Figure 3

(B). Both Gated Recurrent Units (GRUs)


and Long Short Term Memory (LSTM)

[9] cells were explored. The negligible performance difference between the two cell types [10] prompted us to use the GRU model given its simplicity with respect to the LSTM. We utilized a two-level architecture where each layer contains 32 hidden units. The pixel image was reshaped into 100 3-element sequences—one per color channel—where each element consists of 100 values and fed as sequential input into the RNN. The final output layer is identical to the DNN.

Dataset Count Left Center Right
Training (Simulated) 12972 33.60% 33.85% 32.55%
Validation (Simulated) 3243 35.25% 31.85% 32.90%
Test (Simulated) 4054 32.76% 34.46% 32.78%
Real Test (Real-World) 12000 33.33% 33.33% 33.33%
TABLE I: Dataset Counts and Distribution

Iii-E Training: Loss Functions, Optimizers, and Evaluations

All models were trained with the same loss function, optimizer, and evaluation metrics. Cross-entropy was used as the loss function and an Adam Optimizer

[11] with an initial learning rate of 0.001 was used to minimize the cross-entropy loss. The models were evaluated on their accuracy scores, defined as the ratio between the number of correctly labeled images to the total number of images in the set.

Iii-F Neural Network Integration with Unity

In addition to the validation and test accuracy evaluations, and similar to the qualitative evaluation of [4], we devised an evaluation within the Unity environment where the neural network was utilized as a controller of a virtual robot. We instantiated the virtual robot onto one of the virtual trails that was not used during training data collection, ensuring it would not see data it had already been trained on. We then allowed the virtual robot to freely explore the environment, and we qualitatively analyzed its behavior at a high level, seeking to observe whether it was able to navigate the trails or deviate from the trail and wander off into the forest.

Fig. 3: We explored DNN (A), RNN (B), and CNN (C) models for classifying virtual trail imagery.

To accomplish this, the Unity scene was adjusted to allow for the direct control of the virtual robot by the neural network’s classification predictions. In contrast to the virtual data collection where the three-camera paradigm was utilized, a single camera component was positioned at the center orientation of the virtual robot and set to capture images at 30 FPS. This reflects real-world scenarios where robots typically have a single, forward-facing camera. As soon as the image was captured (Figure 4, Step 1), the neural network processed the image (Figure 4, Step 2) and transmitted the classification prediction via UDP socket back to the Unity scene (Figure 4, Step 3). The UDP packet is then parsed within the Unity environment and the virtual robot then moves deterministically based on the neural network prediction.

The control system was designed as follows: a center prediction moves the robot straight ahead, a left prediction slows the robot’s forward movement and rotates it right, and a right prediction slows the robot’s forward movement and rotates it left. The classifications on an image corresponds to the source camera orientation. Consequently, images obtained from the left camera during training contained trails on the right-hand side of the image; as a result, when the model predicts left, the proper control response is to turn right, toward the direction of the trail. Through this pipeline, the virtual robot was set to navigate the virtual path based solely on the neural network’s output of an image taken from a single, forward-facing virtual camera in real time.

Iii-G Evaluation on Real-World Data

The real-world dataset from [4] was utilized as a test set on the models trained on virtual data to demonstrate the feasibility of virtual-to-real-world transfer learning. The set was randomly generated by sampling 4,000 images from each classification (left, center, and right), resulting in 12,000 real-world images. This approach guarantees class balance and establishes the test set baseline at 33% (see Table LABEL:table:data_dist, Real Test). The test set images are processed utilizing the methods described in §III-C, and then fed forward through the virtually trained models to generate a prediction on the real-world image. For every image, the prediction is compared to the image’s true label to obtain the accuracy over the set.

Fig. 4: Step 1: A virtual robot placed within the Unity environment captured images for classification by the neural network. Step 2: the neural network receives the image and produces a probability output. Step 3: classification probabilities are visualized, with the resulting command generated by the maximum class probability sent via UDP back into the Unity environment to control the virtual robot’s movements.
Virtual Test Set Accuracy 88.70% 93.82% 95.02%
Real-World Test Set Accuracy 58.41% 38.60% 48.51%
TABLE II: Model Results

Iv Results

All models were trained and evaluated on virtual data ( pixel images acquired from the Unity environment outlined in §III-A

) for 100 epochs with batch sizes of 128 images. The datasets did not exhibit any significant class imbalance (see Table


); the predominant class of the three was utilized as the baseline to establish whether the models were achieving better results than a policy of continually guessing the majority class. The baseline for the virtual dataset is 35.25%, established by the maximum class imbalance from the validation set; the baseline for the real-world dataset is 33.33%. The models were trained using backpropagation for 50 epochs, which required 1h:23m, 9h:12m, and 2h:4m for the DNN, CNN, and RNN, respectively, on a Macbook Pro with an Intel Iris Pro 1536 MB integrated graphics processor.

Iv-a Virtual Dataset Results

The RNN provided the best results on the virtual dataset scoring a 95.02% on test set accuracy, whereas the DNN provided the best results on the real-world datasets, scoring 58.41%. All three models scored higher than dataset baselines in both virtual and real-world evaluations. The summary of the model performances can be found in Table LABEL:table:model_performance.

Iv-B Unity Follow-Up Evaluations

As mentioned in §III-F, the neural network was integrated with Unity and used as control system for the virtual robot. The RNN model was chosen as the controller due to its top performance on the virtual dataset. In our experiments, we selected sufficiently complex trails—e.g., no straight, level trails—and ensured that the selected trail was not the one used to gather the training data. This ensures that the model is capable of generalizing to novel domains. After selecting an appropriate trail, we placed the virtual robot into the scene and allowed the RNN to govern the autonomous exploration of the environment (see included video submission). Overall, we observed that the robot was largely successful in navigating trails, including those with tight turns and obstacles such as large boulders. Moreover, we observed several instances of “intelligent” decision-making; in one trial, the robot briefly navigates off the trail after colliding with a large obstruction, but then navigates back to the trail and resumes its travel. While promising, we did observe occasional failures. For example, particular terrain regions that exhibited trail-like features, such as small ridgelines, caused the robot to navigate off the trail and begin following these pseudo-trails features.

Iv-C Real-World Dataset Evaluations

Real-world evaluation was conduced on 12,000 images from the real-world dataset described in §III-G. The DNN, CNN, and RNN models achieved classification accuracies of 58.41%, 38.60%, and 48.51%, respectively. Conventional computer vision approaches, such as hue-based saliency mapping coupled with an RBF kernel SVM classifier trained on the real-world dataset comprising our test set have achieved 52.3% clasification accuracies [12, 4]. Significantly, although none of our models achieved the DNN model or human baseline accuracies from [4], our work demonstrates that DNNs trained strictly on virtual data can outperform conventional models trained on real-world data.

V Discussion

The experiments on the virtual datasets demonstrate that the deep learning architectures were capable of learning the correct classifications of virtual images, indicated by the high accuracies, ranging from 88.7% to 95.02%. These scores strongly exceed the data set baselines of non-intelligently predicting the most frequent class. Importantly, the experiments on the real-world images resulted in classification accuracies ranging from 38.60% up to 58.41%, which all exceed the data set baseline of 33.33%. Interestingly, although the virtually trained models did not outperform the CNN or human baselines for real-world test sets, the DNN did outperform the saliency map / SVM baseline from [4] by more than 6%. This suggests that virtual-to-real-world transfer learning utilizing deep learning models may outperform conventional computer vision methods for trail perception. Together, these results indicate that discriminating features for perception of real-world trails have been successfully learned exclusively from virtual trails.

We believe there are several ways to further increase the performance of our virtual-to-real world transfer approach. When conducting pilot tests to iterate over potential network models, we found that longer training periods often ended up reducing real-world test accuracy, suggesting that the models are overfitting on the virtual datasets and would benefit from regularization and training on larger datasets. We suspect that introducing dropout [13] for regularization will yield potentially significant improvements in test set accuracy on the virtual dataset. As an alternative regularization technique, we propose a virtual-real-world fusion data approach for training the models. Specifically, a batch of real-world data could be introduced every virtual-data batches. This approach will likely yield a considerable increase to real-world test set performance, as well as provide a feasible mechanism for bootstrapping real-world robotics systems that utilize deep learning methods for perception, planning, and navigation. In this paradigm of fusion training, only minimal real-world data would need to be collected, with the majority of the training coming from simulations. Conceptually, the models would learn rough approximations in the simulations, and refine important discriminating features via the interspersed real-world training batches.

Interestingly, the RNN outperformed the other models on the virtual dataset, and the evaluations of the models on the real-world dataset yielded counter-intuitive results. Predicated on a suspicion about the sequential order in which images are fed into the RNN, we ran a follow-up experiment wherein the RNN read the images from bottom-to-top as opposed to top-to-bottom. The performance of the RNN decreased substantially and rarely achieved greater accuracies than 50% on the virtual test set. In general, the majority of salient features for trail perception are located within the bottom two-thirds of the image (i.e., the tips of trees is typically uninformative for discerning direction of a trail). When the image is fed from top-to-bottom into the RNN, the information in the top of the image is degraded due to vanishing gradients, which is a well established issue even for LSTM/GRU cell RNN’s. Consequently, when images are fed in bottom-to-top, the most important information is now the first thing the RNN processes and is therefore mostly degraded from the recurrent connections by the end of the image feed.

This result is informative: it is a likely indicator that the classifier is learning to discriminate based on features within the lower half of the image. Consequently, computational demands may be lowered and training made more efficient by training on only bottom half or two-thirds of the image, reducing image processing time and decreasing the number of parameters in the model that must be trained.

To further understand the performance of our models, we analysed incorrectly classified images from the virtual test set. Our analysis points to deficiencies in the models when presented with multiple trails in a single image, suggesting the requirement of a higher level planning system—e.g., GPS and/or compass information of a goal position—to aid the robot’s decision. This analysis also suggests that low quality terrain packs do not allow for sufficient variance amongst objects, obfuscating fine-grained distinctions between trails and other objects with similar features. Consequently, we believe the models may benefit from training on higher quality terrain packs. With state-of-the-art GPUs, virtual environments can be made to closely mimic the appearance of real-life environments and appear nearly photorealistic. We strongly believe virtual scene realism will play a direct role in transfer learning accuracy.

V-a Future Work

Our model was trained using a virtual alpine environment and tested on real data of a similar terrain type. It is likely the model would perform much worse on environments that do not match the synthetic environment’s general terrain characteristics and trail features. Future work will explore procedurally generating terrain with vastly different conditions and features (weather, lighting, biome, path appearance, elevation changes, flora, etc.) to improve generalizability while still being able to rapidly collect large synthetic training datasets. Conveniently, our work allows for rapidly exchanging terrain and environment packages, thus allowing for the development of navigation systems over a large variety of environments and conditions.

One major advantage of our approach is that our data collection process can be automated, drastically increasing the rate of labeled dataset generation. Our current approach captured 20,269 images in less than 5 minutes—a rate of 4,053 images per minute—and is in stark contrast to the 24,474 images collected over a period of 8 hours in [4]. Future work may couple our automated data collection procedure with procedurally generated terrain with higher photorealism to produce additional improvements to this method.

Lastly, an interesting future direction is to discern which features are being learned in the classification task. In a virtual environment, over which we exert complete control, it is possible to filter out one feature at a time, and we can run the same classifier repeatedly in these slightly varied environments. If a feature is turned off and a significant perturbation to classification performance is measured, we can gain insight into the features important for the particular classification task. Running this experiment over numerous terrains may reveal globally important features, enabling us to leverage the statistical properties of these key features for procedurally generated terrain, optimizing the efficiency of the process and enabling more effective results.

Vi Conclusion

In this paper, we trained three different neural network architectures on virtual data generated from Unity and achieved virtual-data classification accuracies ranging from 88% to 95% and real-world classification accuracies ranging from 38% to 58% over a baseline of 33.33%. Robot battery life, human fatigue, and safety considerations present major challenges for manual data collection; however, with our approach, these issues may be circumvented as labeled data generation can be performed rapidly and efficiently within a virtual setting. Robots may then be virtually trained to navigate terrain that is hard to access and/or dangerous, including novel terrains that are currently impossible to access and collect real data from (e.g., Mars) without ever being first exposed to these environments. Our approach demonstrates that virtual-to-real-world transfer learning is a promising approach for overcoming the immense challenges facing real-world data collection and the development of autonomous robotics systems.


This work was supported by a NSF CRII Award #1566612 and an Early Career Faculty grant from NASA’s Space Technology Research Grants Program under award NNX16AR58G. We thank Michael C. Mozer for his help and support of this research.


  • Szafir et al. [2017] D. Szafir, B. Mutlu, and T. Fong, “Designing Planning and Control Interfaces to Support User Collaboration with Flying Robots,” vol. 36, no. 5–7, 2017, pp. 514–542.
  • Peschel and Murphy [2013] J. M. Peschel and R. R. Murphy, “On the human–machine interaction of unmanned aerial system mission specialists,” IEEE Transactions on Human-Machine Systems, vol. 43, no. 1, pp. 53–62, 2013.
  • Hsieh et al. [2007] M. A. Hsieh, A. Cowley, J. F. Keller, L. Chaimowicz, B. Grocholsky, V. Kumar, C. J. Taylor, Y. Endo, R. C. Arkin, B. Jung et al., “Adaptive teams of autonomous aerial and ground robots for situational awareness,” Journal of Field Robotics, vol. 24, no. 11-12, pp. 991–1014, 2007.
  • Giusti et al. [2016] A. Giusti, J. Guzzi, D. C. Cireşan, F.-L. He, J. P. Rodríguez, F. Fontana, M. Faessler, C. Forster, J. Schmidhuber, G. Di Caro et al., “A machine learning approach to visual perception of forest trails for mobile robots,” IEEE Robotics and Automation Letters, vol. 1, no. 2, pp. 661–667, 2016.
  • Yosinski et al. [2014] J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?” in Advances in neural information processing systems, 2014, pp. 3320–3328.
  • Russakovsky et al. [2015] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “ImageNet Large Scale Visual Recognition Challenge,” International Journal of Computer Vision (IJCV), vol. 115, no. 3, pp. 211–252, 2015.
  • Tai et al. [2017]

    L. Tai, G. Paolo, and M. Liu, “Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation,” in

    IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2017, pp. 31–36.
  • Chung et al. [2014] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, “Empirical evaluation of gated recurrent neural networks on sequence modeling,” arXiv preprint arXiv:1412.3555, 2014.
  • Hochreiter and Schmidhuber [1997] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997.
  • Jozefowicz et al. [2015] R. Jozefowicz, W. Zaremba, and I. Sutskever, “An empirical exploration of recurrent network architectures,” in Proceedings of the International Conference on Machine Learning (ICML), 2015, pp. 2342–2350.
  • Kingma and Ba [2014] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  • Santana et al. [2013] P. Santana, L. Correia, R. Mendonça, N. Alves, and J. Barata, “Tracking natural trails with swarm-based visual saliency,” Journal of Field Robotics, vol. 30, no. 1, pp. 64–86, 2013.
  • Srivastava et al. [2014] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” The Journal of Machine Learning Research, vol. 15, no. 1, pp. 1929–1958, 2014.