RoadTagger: Robust Road Attribute Inference with Graph Neural Networks

12/28/2019 ∙ by Songtao He, et al. ∙ MIT Hamad Bin Khalifa University University of Illinois at Urbana-Champaign 9

Inferring road attributes such as lane count and road type from satellite imagery is challenging. Often, due to the occlusion in satellite imagery and the spatial correlation of road attributes, a road attribute at one position on a road may only be apparent when considering far-away segments of the road. Thus, to robustly infer road attributes, the model must integrate scattered information and capture the spatial correlation of features along roads. Existing solutions that rely on image classifiers fail to capture this correlation, resulting in poor accuracy. We find this failure is caused by a fundamental limitation – the limited effective receptive field of image classifiers. To overcome this limitation, we propose RoadTagger, an end-to-end architecture which combines both Convolutional Neural Networks (CNNs) and Graph Neural Networks (GNNs) to infer road attributes. The usage of graph neural networks allows information propagation on the road network graph and eliminates the receptive field limitation of image classifiers. We evaluate RoadTagger on both a large real-world dataset covering 688 km^2 area in 20 U.S. cities and a synthesized micro-dataset. In the evaluation, RoadTagger improves inference accuracy over the CNN image classifier based approaches. RoadTagger also demonstrates strong robustness against different disruptions in the satellite imagery and the ability to learn complicated inductive rules for aggregating scattered information along the road network.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 2

page 4

page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Detailed road attributes enrich maps and enable numerous new applications. For example, mapping the number of lanes on each road makes lane-to-lane navigation possible, where a navigation system informs the driver which lanes will diverge to the correct branch at a junction. Similarly, maps that incorporate the presence of bicycle lanes along each road enable cyclists to make more informed decisions when choosing a route. Additionally, maps with up-to-date road conditions and road types improve the efficiency of road maintenance and disaster relief.

Unfortunately, producing and maintaining digital maps with road attributes is tedious and labor-intensive. In this paper, we show how to automate the inference of road attributes from satellite imagery.

Consider the problem of determining the number of lanes on the road from images. A natural approach would be to map this problem to an image classification problem. Because the number of lanes of one road may vary, we can scan the road with a sliding window and train a classifier to predict the number of lanes in each window along the road independently. After we have the classifier predictions in each window, we can apply a post-processing step to improve the prediction results; e.g., using the road network graph to remove inconsistent predictions along the road. Some prior map inference papers adopt such a strategy [6, 19].

Figure 1: Challenges in road attribute inference. In (a) the lane markings are absent on one side of the road. In (b) the road is occluded by trees and the lane markings are also partially missing. To make correct predictions at all positions on the road, we need to incorporate both the local information and the global information along the road network graph.
Figure 2: Examples of lane inference (examples (a-e)) and road type inference (example (f)). In each image, blue lines show the road graph. The number of lanes or the type of the road predicted by the CNN Image Classifier (with and without MRF) and RoadTagger on each segment are shown along the bottom of each figure. For road type inference, we use capital P to represent primary roads and capital R to represent residential roads. We color the output numbers and letters green for correct predictions and red for incorrect predictions.

This approach suffers from a fundamental limitation: the limited effective-receptive-field of image classifiers. Consider Figure 1(a), where lane markings are visible only on the left. Because the road width remains the same, these lane markings imply the remainder of the road on the right has the same number of lanes, despite not having explicit markings. However, because practical image classifiers can only be scalably trained with small windows of the city-wide satellite imagery as input, a window-based image classifier cannot capture this spatial correlation and would not correctly predict the number of lanes in the right portion of the road. Thus, the limited effective-receptive-field of the classifier does not capture the long-term spatial propagation of the image features needed to accurately infer road attributes.

To overcome this limitation, prior work [17] proposes adding a global inference phase to post-process the output from the local classifiers. We find that this fix is inadequate. For example, see Figure 2(a), where the lane count changes from 4 to 5 near an intersection. The image classifier outputs partially incorrect labels. However, the post-processing strategy (Markov Random Field) cannot fix this problem as the global inference phase only takes the predictions from the image classifier as input and it may not be able to tell whether the number of lanes indeed changes or it is an error of the image classifier. This limitation is caused by the information barrier induced by the separation of local classification and global inference; the global inference phase can only use the image classifier’s prediction as input, but not other important information such as whether trees occlude the road or whether the road width changes.

We propose RoadTagger, an end-to-end road attribute inference framework that eliminates this barrier using a novel combination of a Convolutional Neural Network (CNN) and a Graph Neural Network (GNN) [22]

. It takes both the satellite imagery and the road network graph as input. For each vertex in the road network graph, RoadTagger uses a CNN to derive a feature vector from a window of satellite imagery around the vertex. Then, the information from each vertex is propagated along the road network graph using a GNN. Finally, it produces the road attribute prediction at each vertex. The GNN eliminates the effective-receptive-field limitation of local image classifiers by propagating information along the road network graph. The end-to-end training of the combined CNN and GNN model is the key to the success of the method: RoadTagger doesn’t select features using only the CNN; instead, by backpropagating from the output of the GNN, the information barrier that limited previous post-processing methods is eliminated.

We evaluate the performance and robustness of RoadTagger with both a real-world dataset covering a area in 20 U.S. cities and a synthesized dataset focused on different challenges in road attribute inference. We focus on two types of road attributes: the number of lanes and the type of road (e.g., primary or residential). In the real-world dataset, we show that RoadTagger surpasses a set of CNN-based image classifier baselines (with and without post-processing). Compared with the CNN image classifier baseline, RoadTagger improves the inference accuracy of the number of lanes from 71.8% to 77.2%, and of the road type from 89.1% to 93.1%. This improvement comes with a reduction of the absolute lane detection error of 22.2%. We show output examples in Figure 2. On the synthesized dataset, we found that RoadTagger is able to learn complicated inductive rules and is robust to different disruptions.

2 Related Work

Cadamuro et al. adapt CNN image classifiers to predict road quality from imagery [6]. Najjar et al. use satellite imagery to create a road safety map of a city [19] by adapting a CNN image classifier to assign a safety score to each input satellite image window. Azimi et al. apply a CNN to perform a semantic segmentation for lane markings [3]. However, because these schemes derive labels directly from the CNN, they are only able to infer attributes that pertain to small objects (e.g. a satellite image window or a lane marker), and not attributes over an entire road.

To address this issue, Máttyus et al. propose modeling the problem as an inference problem in a Markov Random Field (MRF) [17]. They develop an MRF model that encodes low-level image features such as edge, pixel intensity, and image homogeneity; high-level image features such as road detector results and car detector results; and domain knowledge such as the smoothness of the road and overlapping constraints. They show that this model can infer various road attributes, including road width, centerline position, and parking lane location. They further extend the approach to take ground images (dashcam) into account [18]. In contrast to this MRF approach, RoadTagger does not require specification of different features and domain knowledge for each road attribute. Instead, RoadTagger extracts the useful image features and domain knowledge through end-to-end learning, making it a powerful and easy-to-use framework for different road attributes.

Recent work has also explored using satellite imagery for fully automated road network inference and building footprint mapping. DeepRoadMapper [16]

segments the satellite imagery to classify each pixel as road or not road. It then extracts the road network from the segmentation result and applies additional post-processing heuristics to enhance the road network. RoadTracer 

[4] trains a CNN model to predict the direction of the road. It starts from a known location on the road and traces the road network based on the direction prediction from the CNN model. Hamaguchi et al. [13] use an ensemble of size-specific CNN building detectors to produce accurate building footprints for a wide range of building sizes.

Figure 3: The overview of RoadTagger road attribute inference framework.

3 RoadTagger

RoadTagger uses both a CNN and a Gated Graph Neural Network (GGNN) [15] to infer road attributes from satellite imagery and the corresponding road network graph. We assume here the road network graph is already available and accurate, since prior work shows how to infer the graph from satellite imagery [4, 16] or GPS traces [5, 2, 10, 9, 7, 21, 14].

Figure 3 shows an overview of RoadTagger. The first step is to densify the road network graph so that there is one vertex every 20 meters. Then for each vertex in the graph, get the satellite imagery in the area around the vertex to a CNN encoder, and rotate it so that the road direction is always vertical in the image. Each image is pixels, corresponding to a meter tile at 12.5 cm/pixel image resolution. This high resolution is needed to capture details on the road such as lane markings. The CNN encoder uses 12 convolutional layers and 3 fully-connected layers to extract a 64-dimension embedding for each vertex. The method then passes the embeddings of all the vertices to the GNN module. The GNN propagates local information around each vertex to neighbouring vertices on the road network graph. After a few steps of information exchanges, the GNN produces the final prediction of each vertex through three additional fully-connected layers and a soft-max layer.

The model can be expressed as

(1)

where is the input satellite image tile at vertex , is the CNN encoder, is the densified road network graph, is the graph neural network module and is the set of output road attribute labels (soft-max) at vertex . We train the model end-to-end with cross-entropy loss using known ground-truth labels at each vertex .

3.1 Graph Neural Network Module

A Gated GNN is an extension of a Recurrent Neural Network (RNN) on a graph, using a Gated Recurrent Unit (GRU) 

[8] to propagate information between adjacent vertices of the graph. We represent the embedding of the vertex as . Before applying the GGNN, we extend the dimension of from 64 to 128 through two fully-connected layers (the function). Extending the dimension of the original embedding helps ensure that we don’t induce an information bottleneck when the information is propagating on the graph. We represent the hidden state at propagation step of vertex as . Then, the basic propagation model on the road network graph can be written as

(2)

Here, is a function representing the set of all logical neighbors of (explained below), is a fully-connected layer, and is a GRU. The algorithm uses this propagation function to propagate the information on the graph for steps, finally producing the prediction labels through three additional fully-connected layers and a soft-max layer.

Graph Structures

In the propagation model (2), the choice of edge placement defined by is critical to the inference performance; because it controls how the vertices in the graph communicate with each other.

We investigate different graph structures derived from the original road network graph to define . These graph structures share the same set of vertices as the original road network graph but may have different edges defined by . Because the graph structure controls the propagation of information, we can use it to restrict communication to only a subset of the graph or enable communication between two vertices that were not connected in the original road network graph. A good graph structure can improve the performance of the graph neural network. We now discuss four example graph structures:

  1. Original Graph. The structure enables communication between all connected vertices in the road network graph. It allows the GNN to learn the best way of communication without any restriction or preference. However, this freedom also makes it less efficient to learn certain types of road attributes; such as for road-specific attributes, when two roads with different road-specific attributes interact at an intersection, the graph neural network model need to make sure the information from one road won’t mess up with the information from the other road.

  2. Road Extraction Graph. This structure helps propagate messages only within the same road. To automate this process, extract road chains belonging to the same logical road from the original graph. Here, we call a sequence of connected edges a “road chain belonging to the same logical road” when the directional difference between any two consecutive edges is less than 60 degrees. We show examples of road chains with different colors in Figure 3. We find that this restriction is helpful for the propagation of road-specific information as it removes the ambiguity at intersections.

  3. Road Extraction Graph (Directional). We can decompose one road chain into two chains with opposing directions. This decomposition yields two separate graph structures. By providing two graph structures with opposite edge direction to the GNN, we can explicitly specify the source of each message. This modification can help the GNN learn more efficiently.

  4. Auxiliary Graph for Parallel Road Pairs. The original graph adapts two separate vertex chains to represent a parallel road pair. This representation prevents communication between the two roads in a parallel road pair. We enable this communication by adding auxiliary edges between the two roads in a parallel road pair. We show an example of the auxiliary edges in Figure 3. The intuition here is that roads belonging to the same parallel road pair often share the same road attributes, e.g., road type. If one road in a pair of parallel roads is occluded by buildings or trees in the satellite image, RoadTagger can still infer the road attributes correctly by incorporating information from the other road through the auxiliary edges.

We find all these graph structures improve RoadTagger in different ways. Thus, we extend the propagation model (2) to support multiple graph structures. Instead of aggregating messages from different graph structures together, we treat them separately. To support different graph structures, we extend the dimension of the hidden state from to , where the messages from the -th graph structure are stored in the -th -dimensional chunk in the hidden state vector. Although messages from different graph structures are stored separately, they can still interact with each other in the GRU ().

3.2 Training RoadTagger

Training a deep model with both a CNN and a GNN is not straightforward. We found that training RoadTagger with standard cross-entropy loss and common anti-overfitting techniques such as data augmentation and dropout was insufficient. This is because, compared with CNN-based image classifiers on the same training dataset, RoadTagger usually has more parameters, but perceives less diversity.

Moreover, nodes in RoadTagger have a larger set of inputs than the local information in CNNs due to the usage of graph neural networks. This extra information makes RoadTagger even easier to overfit. Unfortunately, there is no well-known regularization mechanism that can be applied during training procedure to prevent RoadTagger from overfitting to this extra information.

In order to prevent RoadTagger from overfitting, we explore two training techniques, random vertex dropout and graph Laplace regularization [20].

Random Vertex Dropout. We represent the embedding of vertex as . During training, we randomly pick up 10% of vertices and set their embedding to a random vector where,

(3)

Here, we use to denote element-wise multiplication. We stop the gradient back-propagation for the dropped vertices. This random vertex dropout is an extension of the standard dropout. However, instead of setting the embeddings to all zeros, we set them to a random vector. This is because an all-zeros vector is too easy for the neural network to distinguish. We aim to use this random vertex dropout to simulate scenarios where the road is partially occluded by trees, buildings or bridges. This can increase the diversity of the training dataset and thus reduce over-fitting.

Graph Laplace Regularization.

As we mentioned, there is no regulation mechanism in our training procedure to prevent RoadTagger from abusing the extra information. To overcome this limitation, we add a regularization term based on graph Laplace regularization to the final loss function. We represent the final soft-max output for vertex

as a -dimensional vector , where is the number of classes of the road attribute. Then, the regularization term for vertex can be written as,

(4)

where is the weight of the regularization term at different vertices. We set to zero if vertex and its neighbours have inconsistent ground truth labels. Otherwise, we set to a constant weight factor.

This additional term forces RoadTagger to generate consistent labels for neighbouring vertices regardless of their correctness. It acts as a regularization term for the cross-entropy loss, which only focuses on per-vertex correctness; thus, it reduces overfitting.

4 Evaluation

In this section, we evaluate the performance and robustness of RoadTagger. We compare RoadTagger against CNN image classifier based solutions. In the evaluation, we focus on the architecture comparison between RoadTagger’s end-to-end CNN+GNN framework and the CNN only image classifier solution. We use the same configuration for all the convolutional layers and fully connected layers except the last one in both RoadTagger’s CNN encoder and the CNN image classifier. We also use the same input satellite image size for both RoadTagger and the CNN image classifier.

In the evaluation, we demonstrate RoadTagger’s performance improvement in road attribute inference in a large scale real world environment. We evaluated the performance of different variants of RoadTagger . In addition, we show when and why RoadTagger can yield better performance and analyze its limitations via a robustness study on a synthesized dataset.

4.1 Dataset

We conduct our evaluation on two datasets, one real-world dataset and one synthetic micro-benchmark. For the real-world dataset, we collect the road attributes (ground truth labels) from OpenStreetMap [12] and the corresponding satellite imagery through the Google static map API [11]. This dataset covers 688 area in 20 U.S. cities. We manually verified the labels of 16 of the dataset from four representative cities: Boston, Chicago, Washington D.C., and Seattle. We use one third of the verified dataset as validation dataset and two thirds of it as testing dataset. We use all the remaining dataset as training dataset.

We focus on inferring two types of road attribute: the number of lanes and the types of roads (residential roads or primary roads). We use these two types of road attributes as representatives because they both have spatial correlation such that nearby segments tend to have the same labels.

4.2 Implementation Details

We implemented both RoadTagger and the CNN image classifier using Tensorflow 

[1]. We use both input image augmentation and dropout to reduce over-fitting. For RoadTagger, we set the graph Laplace regularization weight to

and set the propagation step to 8 in the graph neural network. We train the model to predict both the number of lanes and the type of the road simultaneously. We train the model on a V100 GPU for 300k iterations with a learning rate starting from 0.0001 and decreasing by 3x every 30k iterations. For both models, we use a batch size of 128. In RoadTagger, at each iteration, we pick up a random vertex in the road network graph and start a DFS or BFS from it. We use the first 256 vertices in the search result to generate a sub-graph as input graph to RoadTagger. We only consider a random 128 vertices from the 256 vertices in the loss function. For both models, we use batch normalization to speed up training. We use the model that performs best on the validation set as the final model.

4.3 Baselines

We compare RoadTagger against four different baselines, including (1) using only CNN image classifier, (2) using CNN image classifier with smoothing post-processing, (3) using CNN image classifier with Markov Random Field (MRF) post-processing, and (4) using CNN image classifier with larger receptive fields (1.5x and 2.0x).

In the smoothing post-processing approach, we set the probability outputs of each vertex to be the average probability of itself and its neighbouring vertices in the road network graph. This simple post-processing step can remove scattered errors and make the output labels more consistent.

In the MRF post-processing approach, we use a pairwise term in the energy function of MRF to encourage the road segments, which are connected and belonging to the same logical road, to have the same label. The energy function of the post-processing MRF is,

(5)

We find the best hyper-parameters of MRF ( and ) through brute-force search on the validation set. At the inference time, we use belief propagation to minimize the energy function .

We also evaluate the CNN image classifier with larger receptive fields. The original CNN receptive field at each vertex is a 48x48 meter tile. We derive new baseline approaches by enlarging the receptive field to a 72x72 meter tile (1.5x) and a 96x96 meter tile (2.0x).

4.4 Evaluation on Real-World Dataset

Schemes # of Lane Acc. Gain Road Type Acc. Gain ALE Reduction
CNN Image Classifier (naive baseline) 71.8% - 89.1% - 0.374 -
- with smoothing post-processing 74.1% 2.3% 90.6% 1.5% 0.337 9.8%
- with MRF post-processing 73.7% 1.9% 92.2% 3.1% 0.355 5.1%
CNN Image Classifier (1.5x receptive field) 71.8% 0.0% 90.1% 1.0% 0.367 1.9%
- with smoothing post-processing 74.0% 2.2% 91.1% 2.0% 0.340 9.1%
- with MRF post-processing 74.1% 2.3% 92.9% 3.8% 0.340 9.1%
CNN Image Classifier (2.0x receptive field) 68.8% -2.0% 89.1% 0.0% 0.393 -5.1%
- with smoothing post-processing 70.6% -1.2% 89.9% 0.8% 0.371 0.8%
- with MRF post-processing 70.2% -1.6% 91.6% 2.5% 0.386 -3.2%
RoadTagger (ours) 77.2% 5.4% 93.1% 4.0% 0.291 22.2%
Table 1: Performance of RoadTagger and different CNN image classifier baselines. In the table, we highlight both the best and the second best results.

We use the overall accuracy as metrics for both the number of lane prediction and the road type prediction. For the number of lanes prediction, a two-lane road may be incorrectly recognized as a three-lane road or even a six-lane road. However, the overall accuracy metric doesn’t penalize more for the wrong prediction with six lanes than the wrong prediction with three lanes. Thus, we use an additional metric, the absolute lane error (ALE), to take the degree of error into account. We represent the output prediction of vertex as and the corresponding ground truth is , where both and are integers between and . Then, the ALE is defined as,

(6)

We use this absolute lane error (ALE) as a complement to the overall accuracy metric in our evaluation.

Comparison against baselines

We report the overall accuracy of the two types of road attributes and the absolute lane error for the CNN image classifier baselines and RoadTagger in table 1. We show the result of RoadTagger with its best configuration in this table. As shown in the table, RoadTagger surpasses all the CNN image classifier based baselines. Compared with the baseline using only CNN image classifier, RoadTagger improves the inference accuracy of the number of lanes from 71.8% to 77.2%, and of the road type from 89.1% to 93.1%. This improvement comes with a reduction of the absolute lane detection error of 22.2%. Compared with the best baseline (with MRF post-processing and 1.5x larger receptive field), RoadTagger still improves the accuracy of the lane count inference by 3.1 points, which comes with a reduction of the absolute lane detection error of 14.4%, and achieves similar accuracy in road type inference.

We show output examples of the number of lane prediction and the road type prediction in Figure 2. We find RoadTagger performs more robust in many challenging places than the CNN image classifier. This is because the usage of the graph neural network enables RoadTagger to transitively incorporate information from nearby road segments. Meanwhile, during training, unlike the CNN image classifier, RoadTagger treats all vertices on the sub-graph as a whole rather than treating each vertex independently. This doesn’t force RoadTagger to learn how to map the image of a building into a two-lane road when the building occludes the road in the training dataset. Instead, RoadTagger can learn a more generic inductive rule to understand the spatial correlations and effect of different visual features (e.g., bridges, trees, intersections, etc) on road attributes. Although post-processing approach can be applied to fix some of the scattered errors, e.g., example (e) and (f) in Figure 2, it cannot fix errors which requires more context information such as the errors in examples (a-d) in Figure 2. This limitation is due to the information barrier induced by the separation of local classification and global inference.

Comparison within RoadTagger

Scheme # of Lane Road Type ALE
RoadTagger with - - -
- Raw 74.0% 91.2% 0.332
- Road 75.5% 92.3% 0.327
- Road(D) 75.6% 92.0% 0.324
- Raw+Road(D)+Aux 77.2% 93.1% 0.291
Table 2: Impact of different graph structures used in RoadTagger. Here, we use abbreviations to denote different graphs. We use Raw for the original road network graph, Road for the road extraction graph, Road(D) for the road extraction graph with directional decomposition and Aux for the auxiliary graph for parallel roads.

Within RoadTagger, we first compare the performance of RoadTagger with different graph structures. We show results of RoadTagger with different graph structures in Table 2.

For a single graph structure, we find adding more restrictions into the graph structure can yield better performance, e.g., the performance of using road extraction graph is better than the performance of using the original raw road network graph. This is because in the road extraction graph, message propagation in the graph neural network is restricted to be within each logical road. This can remove the ambiguity of message propagation at intersections in the original road network graph, thus, improve performance.

RoadTagger supports using multiple graph structures. We find using the combination of the Raw graph, Road(D) graph and Aux graph can yield better performance compared with the performance of using a single graph structure. This is because using multiple graph structures allows our neural network model to learn the best message propagation graph(s) for different attributes end-to-end.

As we mentioned before, we adopt two training techniques to improve the performance of RoadTagger. We show the comparison results in Table 3. We find both of these two techniques are critical to the performance improvement of RoadTagger; the random vertex dropout has more impact on the number of lane inference and the graph Laplace regularization has more impact on the road type inference.

Scheme # of Lane Road Type ALE
RoadTagger 77.2% 93.1% 0.291
No Vertex Dropout 74.7% 92.7% 0.325
No Regularization. 76.5% 90.8% 0.300
Table 3: Impact of random vertex dropout and graph Laplace regularization.

4.5 Evaluation on Synthesized Micro-Benchmark

Figure 4: Two representative samples of the micro benchmark. RoadTagger predicts both of them correctly.

We conduct an extensive evaluation of RoadTagger on a micro benchmark. In this micro benchmark, we inject different types of challenges to the satellite imagery. We would like to study the impact of occlusions with different types and amounts as well as other challenges such as missing lane markings in a controlled way. In this micro benchmark, we find RoadTagger is robust to a wide range of different disruptions. These disruptions include removing all the lane markings on part of the roads, alternatively occluding the left and right side of the roads, and even occluding the target road with an overpass road.

We show two representative examples of this benchmark in Figure 4. Please refer to the supplementary material for the evaluation on the whole micro-benchmark.

We find the two examples shown in Figure 4 particularly interesting. In example (a), an overpass road occluded the target road. In example (b), the lane count changes when the road is occluded by trees. To correctly predict the number of lanes in both example (a) and (b), RoadTagger needs to know that when the starting point of an overpass road is detected, the visual features of the overpass should be ignored until the far edge of the overpass is detected. At the same time, RoadTagger needs to know if the road is temporarily occluded by trees, the following road segment still belongs to the same target road. We find RoadTagger’s end-to-end architecture enables it to learn all this knowledge correctly without any additional labels or explicit features. This is perhaps the most attractive part of RoadTagger.

5 Discussion

Can RoadTagger Generalize to City-Scale Graphs? In our evaluation, we find RoadTagger can generalize well to city-scale graphs. During inference, RoadTagger labels the whole road network graph (with 3,000 to 4,000 vertices) for each 2km by 2km region in one shot (Use the entire road network graph as input). We find training RoadTagger with 256-node subgraphs can generalize well in larger graphs, e.g., graphs with 3,000 to 4,000 nodes.

Errors Made by RoadTagger. We observe two types of errors made by RoadTagger in our evaluation. (1) We find RoadTagger makes wrong predictions for invisible roads (occluded by trees or buildings) when the disruptions are longer than the GNN propagation step (we show examples of this type of failure in the supplementary material). We think this type of error can be eliminated through enlarging the propagation step and the subgraph size (i.e., 256) during training. (2) We find RoadTagger outputs road attributes in ABABA style along the road when the road attribute is ambiguous. We think this issue can be addressed by incorporating GAN into RoadTagger framework.

6 Conclusion

In this paper, we propose RoadTagger. RoadTagger adapts a novel combination of CNN and graph neural network to enable end-to-end training for road attribute inference. This framework eliminates fundamental limitations of the current state-of-the-art that relies on a single CNN with post-processing. We conduct a comprehensive evaluation of the performance and robustness of RoadTagger; the evaluation result shows a significant improvement in both performance and robustness compared with the current state-of-the-art. The result also shows RoadTagger’s strong inductive reasoning ability learned end-to-end. We believe RoadTagger framework is a fundamental improvement in road attributes inference and can be easily extended to other road attributes.

References

  • [1] M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, et al. (2016)

    Tensorflow: a system for large-scale machine learning

    .
    In 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), pp. 265–283. Cited by: §4.2.
  • [2] M. Ahmed, S. Karagiorgou, D. Pfoser, and C. Wenk (2015) A comparison and evaluation of map construction algorithms using vehicle tracking data. GeoInformatica 19 (3), pp. 601–632. Cited by: §3.
  • [3] S. M. Azimi, P. Fischer, M. Körner, and P. Reinartz (2018) Aerial lanenet: lane marking semantic segmentation in aerial imagery using wavelet-enhanced cost-sensitive symmetric fully convolutional neural networks. arXiv preprint arXiv:1803.06904. Cited by: §2.
  • [4] F. Bastani, S. He, S. Abbar, M. Alizadeh, H. Balakrishnan, S. Chawla, S. Madden, and D. DeWitt (2018) Roadtracer: automatic extraction of road networks from aerial images. In

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    ,
    pp. 4720–4728. Cited by: §2, §3.
  • [5] J. Biagioni and J. Eriksson (2012) Map inference in the face of noise and disparity. In ACM SIGSPATIAL 2012, Cited by: §3.
  • [6] G. Cadamuro, A. Muhebwa, and J. Taneja (2018) Assigning a grade: accurate measurement of road quality using satellite imagery. arXiv preprint arXiv:1812.01699. Cited by: §1, §2.
  • [7] L. Cao and J. Krumm (2009) From GPS traces to a routable road map. In ACM SIGSPATIAL, pp. 3–12. Cited by: §3.
  • [8] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio (2015) Gated feedback recurrent neural networks. In International Conference on Machine Learning, pp. 2067–2075. Cited by: §3.1.
  • [9] J. J. Davies, A. R. Beresford, and A. Hopper (2006) Scalable, distributed, real-time map generation. IEEE Pervasive Computing 5 (4). Cited by: §3.
  • [10] S. Edelkamp and S. Schrödl (2003) Route planning and map inference with global positioning traces. In Computer Science in Perspective, Cited by: §3.
  • [11] Google (2019) Google static maps api. Note: https://developers.google.com/maps/documentation/maps-static/introAccessed: 2019-03-21 Cited by: §4.1.
  • [12] M. Haklay and P. Weber (2008) Openstreetmap: user-generated street maps. IEEE Pervasive Computing 7 (4), pp. 12–18. Cited by: §4.1.
  • [13] R. Hamaguchi and S. Hikosaka (2018) Building detection from satellite imagery using ensemble of size-specific detectors. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 223–2234. Cited by: §2.
  • [14] S. He, F. Bastani, S. Abbar, M. Alizadeh, H. Balakrishnan, S. Chawla, and S. Madden (2018) RoadRunner: improving the precision of road network inference from gps trajectories. In ACM SIGSPATIAL, Cited by: §3.
  • [15] Y. Li, D. Tarlow, M. Brockschmidt, and R. Zemel (2015) Gated graph sequence neural networks. arXiv preprint arXiv:1511.05493. Cited by: §3.
  • [16] G. Máttyus, W. Luo, and R. Urtasun (2017) Deeproadmapper: extracting road topology from aerial images. In Proceedings of the IEEE International Conference on Computer Vision, pp. 3438–3446. Cited by: §2, §3.
  • [17] G. Mattyus, S. Wang, S. Fidler, and R. Urtasun (2015) Enhancing road maps by parsing aerial images around the world. In Proceedings of the IEEE International Conference on Computer Vision, pp. 1689–1697. Cited by: §1, §2.
  • [18] G. Máttyus, S. Wang, S. Fidler, and R. Urtasun (2016) Hd maps: fine-grained road segmentation by parsing ground and aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3611–3619. Cited by: §2.
  • [19] A. Najjar, S. Kaneko, and Y. Miyanaga (2017) Combining satellite imagery and open data to map road safety. In Thirty-First AAAI Conference on Artificial Intelligence, Cited by: §1, §2.
  • [20] A. J. Smola and R. Kondor (2003) Kernels and regularization on graphs. In Learning theory and kernel machines, pp. 144–158. Cited by: §3.2.
  • [21] R. Stanojevic, Abbar,Sofiane, Thirumuruganathan,Saravanan, Chawla,Sanjay, F. Filali, and A. Aleimat (2018) Robust road map inference through network alignment of trajectories. In Proceedings of the 2018 SIAM International Conference on Data Mining, Cited by: §3.
  • [22] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and P. S. Yu (2019) A comprehensive survey on graph neural networks. arXiv preprint arXiv:1901.00596. Cited by: §1.