Updating Street Maps using
Changes Detected in Satellite Imagery
Abstract.
Accurately maintaining digital street maps is labor-intensive. To address this challenge, much work has studied automatically processing geospatial data sources such as GPS trajectories and satellite images to reduce the cost of maintaining digital maps. An end-to-end map update system would first process geospatial data sources to extract insights, and second leverage those insights to update and improve the map. However, prior work largely focuses on the first step of this pipeline: these map extraction methods infer road networks from scratch given geospatial data sources (in effect creating entirely new maps), but do not address the second step of leveraging this extracted information to update the existing digital map data. In this paper, we first explain why current map extraction techniques yield low accuracy when extended to update existing maps. We then propose a novel method that leverages the progression of satellite imagery over time to substantially improve accuracy. Our approach first compares satellite images captured at different times to identify portions of the physical road network that have visibly changed, and then updates the existing map accordingly. We show that our change-based approach reduces map update error rates four-fold.
1. Introduction
Maintaining street maps is a labor-intensive process. As a result, many techniques have been proposed to automate parts of this process by using geospatial data sources. Current map extraction techniques (Alshehhi et al., 2017; Bastani et al., 2018b; Cheng et al., 2017; Costea et al., 2017; Máttyus et al., 2017; Panboonyuen et al., 2017; Vakalopoulou et al., 2015) primarily rely on satellite imagery due to its global availability, while some techniques use GPS trajectories.
A key problem with current techniques is that they are designed to infer road networks from scratch — however, given that we already have existing high quality maps that cover the vast majority of the world, these inferred road networks are not directly useful. Instead, an end-to-end map update system must process geospatial data sources to update and improve existing digital maps.
1.1. Map Extraction Methods Perform Poorly on Map Update
We first consider extending current map extraction techniques for updating maps. We will show that these methods perform poorly on map update, creating many false positive updates.
Suppose that the current live digital map has a set of roads . We begin by applying a map extraction method to process the most recent satellite imagery (spanning the world). This method produces another set of roads detected in the satellite imagery. Simply replacing with would not be sensible for several reasons:
-
(1)
For roads that appear in both and , given that is largely human-curated, it captures those roads substantially more accurately than .
-
(2)
includes roads such as tunnels that cannot be detected by the map extraction method.
-
(3)
Roads in are labeled with rich annotations such as street names, speed limits, etc. that would be lost in the replacement.
We could instead try to combine and : if for a road segment , and , we add to the map. (We could also remove segments where and , but this would prune roads that are not visible in the satellite image such as tunnels and roads occluded by buildings or trees, so we do not consider it further.)

However, in practice, this approach yields a large number of false positives, where the map extraction method erroneously outputs many road segments in places where there are no roads. We demonstrate this issue by using a state-of-the-art inference method, MAiD (Bastani et al., 2018a), to update OpenStreetMap (Haklay and Weber, 2008). We select a region of Massachusetts where OpenStreetMap has high coverage, and manually remove 204 groups of roads from the map that correspond to new construction between 2015 and 2017. We apply MAiD to recover these removed roads, and score its performance in terms of precision and recall comparing recovered groups of roads to the manually removed groups. At 80% recall, MAiD yields only 67% precision — far too low for full automation to be a realistic option. We show examples of incorrect detections in Figure 1.
Many of these detections, such as the Runway, Walkway, and Crop Field examples in Figure 1, arise due to paths that are virtually indistinguishable from roads: it is restrictions set by policy governing the use of those paths, and not physical characteristics of the paths, that make them unsuitable for traversal by motor vehicles. Thus, simply improving the machine learning techniques and models used in map extraction methods is unlikely to improve accuracy; instead, a fundamental shift in approach is required.
1.2. Key Tasks in Maintaining Maps
To determine how accuracy can be improved, we first take a step back and identify the major challenges associated with maintaining maps. In particular, digital maps have near-complete coverage in most parts of the world: a 2015 study111See https://blog.mapbox.com/how-complete-is-openstreetmap-7c369787af6e. found that, in 154 of 233 considered countries and territories, the length of roads in OpenStreetMap exceeded the estimated road network length reported in the CIA World Factbook, implying that the map provided excellent coverage in those countries. Then, given that digital maps have good coverage, identifying pre-existing roads constructed several years ago is not a key issue: in almost all cases, such roads already appear in the map.
Instead, the key challenge in maintaining maps is keeping the map up-to-date with changes in the physical road network. Indeed, in the US alone, an estimated 30K km of roads are constructed each year222“Public Road Mielage”, FHWA, https://www.fhwa.dot.gov/policyinformation/statistics/2013/vmt422c.cfm., and map vendors spend hundreds of millions of dollars annually to keep maps up-to-date.
This presents the question: Can we develop techniques that directly tackle the key challenge of identifying physical road network changes, to substantially improve accuracy at maintaining maps over map extraction methods?
1.3. Map Update through Change Detection
To substantially improve accuracy, we propose leveraging a source of data largely overlooked in prior work: the progression of satellite imagery over time. By comparing satellite images captured at different times, we can hone in on portions of the road network that visibly changed over the satellite image time series. Focusing on segments that visibly changed over time enables us to disambiguate false positive road segments from genuinely new constructions: for example, all of the false positive detections in Figure 1 arose from curvilinear features such as walkways and crop field paths that did not undergo any recent change; thus, by processing the satellite image time series, we can determine that these are pre-existing features, and should not be added to the map.
To implement the proposed solution, we must detect changed road segments across satellite images. However, most existing change detection methods are fully supervised. They rely on collecting annotated pairs of images where change has occurred. Since newly constructed roads are rare relative to the size of the map, collecting positive examples of new roads for such a dataset is tedious and costly. Additionally, the diversity in visual appearance of roads makes change detection especially challenging. Furthermore, in Section 4, we show that prior work in unsupervised change detection exhibits low accuracy when applied for detecting new construction.
Instead, we develop a two-stage approach that requires no hand-labeling for comparing satellite images over time to detect new roads. In the first stage, we apply a novel change-seeking iterative tracing procedure to detect recently constructed roads that are missing from the existing map. Our method uses ground-truth road labels derived from the existing map dataset to avoid needing new annotations, and detects new roads that appear in an up-to-date satellite image but are not visible in an old image.
Though the first stage is effective at detecting new construction, it nevertheless yields false positive detections when occlusion and other factors yield visual differences between the old and up-to-date images despite no actual change. Thus, in the second stage, we propose a novel self-supervised change detection approach to further improve precision. We train a CNN to classify whether windows of two aligned satellite images captured at different times are cropped at the same window (matching) or at different windows (mismatched). The CNN learns to match features like road markers to determine whether two crops are matching or mismatched. To apply the model for inference, we provide it with matching crop pairs around roads detected through tracing, and we only retain detections that fool the model into classifying the pair as mismatched, suggesting the presence of new construction.
We evaluate our approach on a large-scale dataset consisting of 4800 km2 of satellite imagery. We apply approaches to improve an existing map dataset, OpenStreetMap, by adding newly constructed roads to the map. At 50% recall, our approach reduces error rates over existing state-of-the-art map inference methods four-fold, from 12% to 3%. Our code and data is available at https://favyen.com/mapupdate.
In summary, our contributions are:

-
•
We propose a novel approach for updating street maps that leverages the progression of satellite imagery over time. In contrast to prior work that focuses on inferring all roads in a satellite image, our approach tackles the directly practical map update problem.
-
•
We develop a two-stage approach for detecting new construction in satellite imagery. Our approach does not require any hand-labeling. In the second stage, we propose a novel self-supervised road-masking approach, where we train a CNN classifier in a self-supervised manner to detect change.
-
•
We evaluate our approach on a large-scale dataset consisting of 3000 km2 of satellite imagery in the Boston area for training and 1800 km2 in northeastern Massachusetts for testing. We apply approaches to improve an existing map dataset, OpenStreetMap, by incorporating newly constructed roads into the dataset. At 50% recall, our approach reduces error rates over existing state-of-the-art map inference methods four-fold, from 12% to 3%.
2. Related Work
Road Extraction. Automatically inferring roads from satellite imagery is a well-studied problem. Recent road extraction methods generally apply convolutional neural networks (CNNs) to segment imagery for roads (Batra et al., 2019; Costea et al., 2017; Mosinska et al., 2019; Panboonyuen et al., 2017; Yang et al., 2019; Zhou et al., 2018), and apply various methods to post-process the segmentation output and derive vector road network graphs. Cheng et al. apply binary thresholding, morphological thinning, and line following to extract a road network from segmentation probabilities (Cheng et al., 2017). DeepRoadMapper (Máttyus et al., 2017) proposes several additional heuristic and learning-based refinement steps, including removing short edges and identifying potential missed roads. Some methods propose alternatives to treating road extraction as an image segmentation problem. Alshehhi et al. build the road network with a region adjacent graph that forms narrow elongated regions along roads (Alshehhi et al., 2017). RoadTracer (Bastani et al., 2018b) and PolyMapper (Li et al., 2019) propose an iterative tracing framework to extract road networks: they train a CNN to output the directionality of roads at each pixel, and employ an iterative search guided by the CNN to trace the road network. VecRoad extends the iterative tracing approach with a flexible step size and joint learning tasks (Tan et al., 2020), and Neural Turtle Graphics extends it with a sequential generative model (Chu et al., 2019). Another recent technique, Sat2Graph, proposes a one-shot road extraction process where a CNN directly predicts the positions of road network vertices and edges (He et al., 2020).
However, broadly, these approaches are unable to reason about false positive detections made by the CNN such as those in Figure 1, especially for paths that appear visually similar to roads in the satellite image but are not suitable for traversal by motor vehicles. As a result, when road extraction methods are applied for the practical task of updating existing maps, they incorporate many non-road paths into the map, thereby substantially deteriorating the quality of the map data. In contrast, in most of the world where existing maps have good coverage, our method accurately keeps maps up to date with new construction by comparing satellite imagery over time, and only updating the map in areas where change is detected across images.
GPS Trajectories for Updating Maps. Inferring roads from GPS trajectory data has also been studied (Ahmed et al., 2015; Biagioni and Eriksson, 2012; Cao and Krumm, 2009; Davies et al., 2006; Prabowo et al., 2019; He et al., 2018; Stanojevic et al., 2018). Two works in this space, CrowdAtlas (Wang et al., 2013) and COBWEB (Shan et al., 2015), propose map update methods to incorporate new roads into existing maps. However, since these methods do not consider GPS time series data (comparing older trajectories to recent trajectories), they exhibit false positive errors similar to satellite image road extraction methods due to GPS noise. Additionally, due to the lack of a ground truth test set, prior work have not incorporated a quantitative evaluation of the map update portion of those methods, and instead qualitatively show results at detecting a small number of roads missing from an existing map. While it may be possible to accurately update existing maps by comparing old and recent GPS trajectory data, this has not been studied in prior work; in our approach, we focus on using satellite image time series data, since satellite imagery is globally available.
Change Detection. Change detection in satellite imagery has previously been studied for detecting damage from natural disasters and armed conflict. Gueguen et al. employ a semi-supervised learning approach to identify damaged regions by comparing images before and after a calamity(Gueguen and Hamid, 2015). However, adapting supervised and semi-supervised change detection methods for maintaining maps is difficult: annotating examples of new roads is highly time-consuming because the density of new construction is low. Unsupervised change detection methods have also been proposed (de Jong and Bosman, 2019), but we will show in Section 4 that these methods exhibit poor accuracy when used to identify newly constructed roads.

3. Detecting Street Network Changes
Existing state-of-the-art map inference methods are designed to infer maps from scratch rather than update existing maps. When applied for updating maps with new roads, these methods show poor accuracy — the number of false positives overwhelms the few cases of new construction. Figure 1 shows examples of false positive detections that arise when we apply these methods to add missing roads to OpenStreetMap. Oftentimes, these false positives arise due to pre-existing paths such as air fields and cycling paths that appear visually similar to roads (Figure 1). We propose a novel approach that, in contrast to prior work, directly detects newly constructed roads by comparing an up-to-date satellite image against an old image. Figure 2 summarizes our approach.
Let and be old and up-to-date satellite images of the same region, and let be the road network graph of that region in the existing map. Each vertex is annotated with a pixel corresponding to its location in the images, and edges correspond to roads. At a high level, in our approach, we detect new roads by identifying roads that appear in but not in or . In most of the world where existing maps have good coverage, roads detected in both and likely are not roads at all, but instead non-road paths (e.g., the examples in Figure 1); by comparing and , our approach avoids these false positives.
We first apply a novel change-seeking iterative tracing procedure that adapts MAiD (Bastani et al., 2018a) to selectively trace roads in that appear in neither nor the existing map , i.e., roads that were constructed after was captured. Our method traces roads along segments where a CNN model has high confidence in and low confidence in . Although this procedure improves precision over prior work, we find that it produces false positives when differences in off-nadir angle and lighting or visible non-construction activity result in a sharp increase in the CNN confidence from to despite no new roads.
Thus, we propose a novel self-supervised change detection method to automatically prune these remaining false positives in the second stage of our approach. Our method selectively identifies road network changes so that visible non-construction activity does not result in a false positive. The final road detections have high precision, and can be used to improve real-world maps through automatic merging or human validation.
In Section 3.1, we describe our method to obtain an initial set of candidate roads using change-seeking iterative tracing. We then introduce our novel self-supervised road-masking approach in Section 3.2.
3.1. Change-Seeking Iterative Tracing
In the first stage, we apply a MAiD (Bastani et al., 2018a) model to segment the images for tracing confidences; each tracing confidence indicates the likelihood that a road passes a pixel in a particular direction (angle). In MAiD, these tracing confidences are used by an iterative tracing algorithm to draw roads along directions with high confidence. In contrast, we develop a change-seeking iterative tracing process that avoids many of the false positives in Figure 1 that involve pre-existing non-road paths by comparing confidences extracted from and from to draw roads only along directions with substantially higher confidence in , which suggests the presence of a new road. Below, we detail each of the components in our approach.
Model. We use the MAiD CNN model architecture from (Bastani et al., 2018a). Given an image , the model produces a three-dimensional matrix , where is the probability that a road passes the pixel at in a direction specified by . includes 64 channels, and the th channel indicates the likelihood that there is a road at an angle between and from a pixel. At a pixel that falls along a straight road, channels specifying opposite directions along the road would both have high confidence. is output at one-fourth the input resolution.
Training. During training, we construct an example input by first randomly deciding whether to input or , and then selecting a random 2D window in the image. We create training labels using the existing road network graph . The input is a image (where the three channels are derived from RGB satellite imagery), and the label is a matrix of tracing confidences. We train the model using binary cross entropy loss, averaged across pixels and the 64 channels.
Tracing Procedure. We develop a novel change-seeking iterative tracing procedure that adapts the tracing process used in prior work (Bastani et al., 2018b, a) to focus tracing on newly constructed roads and avoid inferring false positives along pre-existing non-road paths. Figure 3 illustrates the tracing process.
Iterative tracing starts from an initial pixel known to lie on the road network, and follows directions with high confidence in at the current pixel in a depth-first search (DFS) process. The inferred road network consists of the paths that were followed during the search.
We first compute from and from . We use the existing map as a base map: our tracing procedure starts from pixels in the base map, and follows directions with high confidence in and low confidence in in a depth-first search (DFS) process. The inferred road network consists of paths that were followed during the search. Let be the current road network state, which we extend during tracing. We initialize by densifying so that vertices are at most 10 m apart. We use each vertex in as a starting pixel for tracing, and append all vertices to a DFS stack.
On each tracing iteration, we consider the pixel at the top of the search stack. We identify the highest confidence direction in that has an angular distance of at least from any existing edges in at . This prevents re-tracing roads that are already covered by . Additionally, though, we only trace from if and , i.e., if the tracing confidence in the up-to-date imagery exceeds a threshold while the confidence for the same pixel and direction in the old imagery is small. Thus, our tracing procedure only follows new roads that are not reflected in . If we decide to trace, then we add a new vertex to , where is the angle corresponding to , and add an edge from to . We then push onto the DFS stack. Otherwise, if we decide not to trace, we pop from the stack.
We terminate tracing once the DFS stack is empty. At this point, each connected component in is a candidate group of roads. By comparing and during the tracing procedure, we are able to avoid tracing along pre-existing non-road paths.
3.2. Self-Supervised Learning for Selective Change Detection

Change-seeking iterative tracing improves precision over prior work, but still produces false positives when reflects higher confidence than due to angle and lighting differences or visible activity without new construction between and . Figure 4 shows example false positives. On the left, a railway and a pedestrian path are partially occluded by shadows in the old image, but are visible in the up-to-date image. On the right, although there is visible activity (bridge demolition and crosswalk painting), there are no new roads, and two walkways are incorrectly detected.
In the second stage of our approach, we apply change detection to filter the candidates generated in the first stage by pruning these false positives. However, as we will show in the evaluation, unsupervised change detection methods are unable to robustly distinguish false positives due to angle, lighting, and other aforementioned differences. Supervised methods are also impractical: paired examples of new construction are tedious to annotate due to their low density. Instead, we develop a novel approach that applies self-supervised learning to selectively identify road network changes.

In our self-supervised learning procedure, we train a classifier that inputs a pair of windows of old and up-to-date images. The input may either be a matching pair, where and are cropped at the same window, or a mismatched pair, where and are cropped at disjoint windows. We train the classifier to distinguish matching pairs from mismatched pairs. We generate training examples by deciding to create a matching pair or mismatched pair with equal probability. To generate a matching pair, we randomly pick one window and crop both images at that window. To generate a mismatched pair, we randomly pick two disjoint windows. The classifier learns to match features between the images to determine whether they are taken at the same window despite differences in matching pairs such as shadows, camera angle, and non-construction activity that make unsupervised change detection methods ineffective on this task. We show example training pairs in Figure 5.
To apply the filter for inference, we execute the classifier on crops of and taken at a window around connected components of roads detected during tracing. Although this pair is matching, where the crops are aligned, if there is substantial change in the images due to newly constructed roads, the crops are likely to fool the classifier into outputting a higher probability for the “mismatched” class. Thus, we prune candidate roads if the “mismatched” probability falls below a threshold.
However, in practice, new construction is often adjacent to pre-existing roads, buildings, and other structures. If the classifier observes the same structures in both the old and up-to-date windows, it can determine that the pair is matching with high confidence. Thus, we develop a masking approach that focuses the classifier on the detected candidate road.


Model. The model input consists of 7 channels: 3 channels from the crop of , 3 channels from the crop of , and 1 channel containing a mask . is either 0 or 1, and if , then we zero the corresponding values in the crop channels, i.e., . Because the size of the input during inference varies based on the candidate road, we use a fully convolutional CNN architecture consisting of 6 encoder layers followed by 5 decoder layers. Figure 6 shows the model architecture. The model outputs a probability at each pixel that the input example is “matching”. We train the CNN with cross entropy loss, averaged over only pixels where .
Training. On each training step, we construct a matching example with 50% probability and a mismatched example with 50% probability. In both cases, we begin by computing the mask that we will apply to the imagery crop inputs. During inference, will be 1 only near a group of candidate roads obtained through tracing. For effective training, must be similar to what we will provide during inference, e.g., should predominantly cover roads. Thus, we leverage to compute the mask: we randomly select a vertex in , and perform a breadth-first-search from to derive a subgraph that will determine . During the search, we add each traversed edge to , and terminate the search once the length of the bounding box containing exceeds a threshold . We vary to ensure that training examples have diverse mask sizes, since during inference, candidate groups of roads may exhibit different sizes; specifically, we pick uniformly between 50 m and 150 m. We set if pixel falls within 20 meters of some edge in , and set otherwise.

Method | APLS |
---|---|
DeepRoadMapper | 0.52 |
MAiD | 0.49 |
UnstructChange | 0.48 |
CmpOnly | 0.51 |
Cmp+Filter (ours) | 0.57 |

To create a mismatched example, we select two windows and . is always a window centered at . We randomly pick so that and are disjoint, but the distance from to is at most 2500 m. Choosing nearby but disjoint windows when creating mismatched examples is crucial as it yields more challenging examples where the old and up-to-date crops have similar style (e.g., both in suburban neighborhoods) but different semantic content. We crop at , and and at .
To create a matching example, with 80% probability, we simply crop both and at . However, in some cases, tracing may output roads over non-road paths; if we only train the model on matching examples where the pixels where fall on roads, the model may be ineffective on non-road inputs. Thus, with 20% probability, we crop and at a random window .
Figure 5 shows two training examples.
Inference. For each candidate group of roads , we first derive a corresponding mask similar to the process during training: only if pixel falls within 20 meters of a candidate road. We crop , , and at a window corresponding to the bounding box of with 20-meter padding. We then compute the average probability that the model outputs over pixels where . Then, given a filter threshold , if , we prune the candidate. In Figure 7, we show the average probability on several roads inferred through change-seeking iterative tracing.
4. Evaluation
We evaluate our approach against existing state-of-the-art map inference and change detection methods on a task involving automatically updating OpenStreetMap with new roads. We use 60 cm/pixel resolution satellite imagery from MassGIS from 2015 and 2017 as our old imagery and up-to-date imagery , and the OpenStreetMap dataset as our road network graph. We select two disjoint sections of this dataset for training (3000 in the Boston metro area) and for evaluation (1800 in northeastern Massachusetts).
Metrics. For evaluation, we hand-annotated 204 groups of roads that appear in but not . We prune these roads from OpenStreetMap to derive a road network corresponding to a map that has not yet been updated with the new imagery. We compare the methods in terms of the precision and recall on recovering the pruned roads. Each approach outputs a set of map update proposals , where each is a connected component of inferred roads. The pruned roads form a set of ground truth proposals . We say a proposal matches a ground truth proposal if the proposal bounding boxes intersect. Then, precision and recall are defined as:
Under this metric, each approach yields a precision-recall curve over varying confidence thresholds (e.g., in our approach, varying ). Because is not comprehensive, we discard a proposal if it is a correct example of a road but does not match with any ground truth proposal.
Although the focus of our approach is on improving the precision of inferred road segments rather than improving the geometrical accuracy of those segments, we also evaluate the methods on the latter in terms of Average Path Length Similarity (APLS) (Van Etten et al., 2018).
Baselines. We evaluate our method against two baselines detailed in Section 2 that implement existing state-of-the-art road inference approaches: MAiD (Bastani et al., 2018a) and DeepRoadMapper (Máttyus et al., 2017). We apply these methods on to derive proposed roads, and prune proposals that correspond to roads already mapped in OpenStreetMap by pruning segments that fall within 40 m of an edge in . We also evaluate against an unsupervised satellite image change detection method, UnstructChange (de Jong and Bosman, 2019), which identifies change by comparing feature maps extracted from old and up-to-date satellite images through a VGG-19 model trained for segmentation. To apply UnstructChange, we first obtain candidate roads through MAiD, and then eliminate candidates where the unsupervised method detects no change.
Finally, a fourth baseline, denoted CmpOnly, applies the first stage of our approach only (change-seeking iterative tracing). We denote our full approach Cmp+Filter.
Results. We show precision-recall curves on detecting new roads over varying confidence thresholds in Figure 8, and qualitative results in Figure 9. DeepRoadMapper is unable to achieve higher than 91% precision due to false positives, many of which correspond to pre-existing non-road paths. MAiD provides higher precision at lower recalls, but still yields only 88% precision at 50% recall. At 50% recall, CmpOnly improves precision to 94%, and Cmp+Filter further improves precision to 97%. Thus, our method effectively prunes false positives that have lighting and angle differences or visible activity but no new roads. UnstructChange does not improve performance over MAiD: it outputs false positives due to non-construction changes (such as angle and lighting differences) between the old and up-to-date images. While our work focuses on improving the precision of road detections, Figure 8(b) shows that our method also yields a 5% improvement in APLS, which measures the geometrical accuracy of inferred roads.
Overall, our full approach provides near-100% precision at reasonable recall levels. Precision is crucial because automatic integration of detections into the street map dataset is only practical if errors are rare – otherwise, the confusion for users caused by introducing errors may outweigh the benefit from expanded map coverage.
4.1. Updating Maps with New Buildings
Street maps contain numerous annotations besides roads. In this section, we show that a simple adaption of our approach for detecting new construction of buildings yields high accuracy.


Baselines. We evaluate against two baselines. BldgSeg implements segmentation-based building extraction (Li et al., 2018), applying a deep CNN to segment imagery and then extracting building polygons from the segmentation probabilities. BldgSeg Cmp applies the first stage of our approach, which we adapt for buildings below.
First Stage. Our change-seeking iterative tracing method is effective at tracing road networks, but is not applicable for tracing building polygons. Instead, we apply the BldgSeg baseline to segment and for buildings, and derive candidate buildings by comparing segmentation probabilities. Specifically, we first compute segmentation probabilities and using the CNN at each pixel in the imagery. We then compute a binary image such that only if and . Finally, we extract buildings from .
Second Stage. We train and apply our self-supervised model as for roads, but compute the mask by adding a fixed padding around building polygons instead of around roads.
Metrics. We evaluate the methods on precision and recall, as defined for roads. We construct a ground truth set of building polygons by removing 665 buildings from OpenStreetMap that were constructed in northeastern Massachusetts between 2015 and 2017.
Results. We show results in Figure 10. On buildings, at 30% recall, BldgSeg yields only 85% precision. BldgSeg Cmp improves precision to 87%, and our approach, Cmp+Filter, yields 92% precision. This corresponds to an almost two-fold reduction in error rate (), from 15% to 8%.
4.2. Removing Roads

Cmp+Filter can be applied in reverse to identify removed roads, where we identify portions of the existing map that appear in but not in . Although we are unable to identify enough examples of removed roads to conduct a quantitative evaluation, we show three detections of removed roads in Figure 11. Cmp+Filter succeeds in identifying a shifted highway and bulldozed road.
5. Conclusion
Maintaining street maps today is labor-intensive and costly. We find that existing state-of-the-art street map inference systems exhibit low precision when applied to update an existing map dataset, OpenStreetMap. By leveraging multiple satellite images collected at different times, our two-stage approach complements prior work by identifying roads and buildings that were newly constructed in the most recent image. Our evaluation on 4800 of satellite imagery shows that our approach is able to update existing maps to capture new construction with high precision.
References
- (1)
- Ahmed et al. (2015) M. Ahmed, S. Karagiorgou, D. Pfoser, and C. Wenk. 2015. A Comparison and Evaluation of Map Construction Algorithms using Vehicle Tracking Data. GeoInformatica 19, 3 (2015), 601–632.
- Alshehhi et al. (2017) Rasha Alshehhi, Prashanth Reddy Marpu, Wei Lee Woon, and Mauro Dalla Mura. 2017. Simultaneous Extraction of Roads and Buildings in Remote Sensing Imagery with Convolutional Neural Networks. ISPRS Journal of Photogrammetry and Remote Sensing 130 (2017), 139–149.
- Bastani et al. (2018a) Favyen Bastani, Songtao He, Sofiane Abbar, Mohammad Alizadeh, Hari Balakrishnan, Sanjay Chawla, and Sam Madden. 2018a. Machine-Assisted Map Editing. In ACM International Conference on Advances in Geographic Information Systems (SIGSPATIAL). 23–32.
- Bastani et al. (2018b) Favyen Bastani, Songtao He, Sofiane Abbar, Mohammad Alizadeh, Hari Balakrishnan, Sanjay Chawla, Sam Madden, and David DeWitt. 2018b. RoadTracer: Automatic Extraction of Road Networks from Aerial Images. In CVPR.
- Batra et al. (2019) Anil Batra, Suriya Singh, Guan Pang, Saikat Basu, CV Jawahar, and Manohar Paluri. 2019. Improved Road Connectivity by Joint Learning of Orientation and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 10385–10393.
- Biagioni and Eriksson (2012) James Biagioni and Jakob Eriksson. 2012. Map Inference in the Face of Noise and Disparity. In ACM SIGSPATIAL.
- Cao and Krumm (2009) Lili Cao and John Krumm. 2009. From GPS Traces to a Routable Road Map. In ACM SIGSPATIAL. 3–12.
- Cheng et al. (2017) Guangliang Cheng, Ying Wang, Shibiao Xu, Hongzhen Wang, Shiming Xiang, and Chunhong Pan. 2017. Automatic Road Detection and Centerline Extraction via Cascaded End-to-End Convolutional Neural Network. IEEE Transactions on Geoscience and Remote Sensing 55, 6 (2017), 3322–3337.
- Chu et al. (2019) Hang Chu, Daiqing Li, David Acuna, Amlan Kar, Maria Shugrina, Xinkai Wei, Ming-Yu Liu, Antonio Torralba, and Sanja Fidler. 2019. Neural Turtle Graphics for Modeling City Road Layouts. In Proceedings of the IEEE International Conference on Computer Vision (ICCV). 4522–4530.
- Costea et al. (2017) Dragos Costea, Alina Marcu, Emil Slusanschi, and Marius Leordeanu. 2017. Creating Roadmaps in Aerial Images With Generative Adversarial Networks and Smoothing-Based Optimization. In ICCV.
- Davies et al. (2006) Jonathan J Davies, Alastair R Beresford, and Andy Hopper. 2006. Scalable, Distributed, Real-time Map Generation. IEEE Pervasive Computing 5, 4 (2006).
- de Jong and Bosman (2019) Kevin Louis de Jong and Anna Sergeevna Bosman. 2019. Unsupervised Change Detection in Satellite Images Using Convolutional Neural Networks. In International Joint Conference on Neural Networks (IJCNN). IEEE, 1–8.
- Gueguen and Hamid (2015) Lionel Gueguen and Raffay Hamid. 2015. Large-Scale Damage Detection Using Satellite Imagery. In CVPR.
- Haklay and Weber (2008) Mordechai Haklay and Patrick Weber. 2008. OpenStreetMap: User-Generated Street Maps. IEEE Pervasive Computing 7, 4 (2008), 12–18.
- He et al. (2018) Songtao He, Favyen Bastani, Sofiane Abbar, Mohammad Alizadeh, Hari Balakrishnan, Sanjay Chawla, and Sam Madden. 2018. RoadRunner: Improving the Precision of Road Network Inference from GPS Trajectories. In ACM SIGSPATIAL.
- He et al. (2020) Songtao He, Favyen Bastani, Satvat Jagwani, Mohammad Alizadeh, Hari Balakrishnan, Sanjay Chawla, Mohamed M Elshrif, Samuel Madden, and Amin Sadeghi. 2020. Sat2Graph: Road Graph Extraction through Graph-Tensor Encoding. Proceedings of the European Conference on Computer Vision (ECCV) (2020).
- Li et al. (2018) Weijia Li, Conghui He, Jiarui Fang, and Haohuan Fu. 2018. Semantic Segmentation Based Building Extraction Method Using Multi-Source GIS Map Datasets and Satellite Imagery. In CVPR.
- Li et al. (2019) Zuoyue Li, Jan Dirk Wegner, and Aurélien Lucchi. 2019. Topological Map Extraction from Overhead Images. In Proceedings of the IEEE International Conference on Computer Vision (ICCV). 1715–1724.
- Máttyus et al. (2017) Gellért Máttyus, Wenjie Luo, and Raquel Urtasun. 2017. DeepRoadMapper: Extracting Road Topology From Aerial Images. In CVPR.
- Mosinska et al. (2019) Agata Mosinska, Mateusz Koziński, and Pascal Fua. 2019. Joint Segmentation and Path Classification of Curvilinear Structures. IEEE Transactions on Pattern Analysis and Machine Intelligence 42, 6 (2019), 1515–1521.
- Panboonyuen et al. (2017) Teerapong Panboonyuen, Kulsawasd Jitkajornwanich, Siam Lawawirojwong, Panu Srestasathiern, and Peerapon Vateekul. 2017. Road Segmentation of Remotely-Sensed Images using Deep Convolutional Neural Networks with Landscape Metrics and Conditional Random Fields. Remote Sensing 9, 7 (2017), 680.
- Prabowo et al. (2019) Arian Prabowo, Piotr Koniusz, Wei Shao, and Flora D Salim. 2019. COLTRANE: Convolutional Trajectory Network for Deep Map Inference. In Proceedings of the ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation. 21–30.
- Shan et al. (2015) Zhangqing Shan, Hao Wu, Weiwei Sun, and Baihua Zheng. 2015. COBWEB: A Robust Map Update System using GPS Trajectories. In Proceedings of the ACM International Joint Conference on Pervasive and Ubiquitous Computing. 927–937.
- Stanojevic et al. (2018) Rade Stanojevic, Sofiane Abbar, Saravanan Thirumuruganathan, Sanjay Chawla, Fethi Filali, and Ahid Aleimat. 2018. Robust Road Map Inference through Network Alignment of Trajectories. In Proceedings of the 2018 SIAM International Conference on Data Mining. SIAM.
- Tan et al. (2020) Yong-Qiang Tan, Shang-Hua Gao, Xuan-Yi Li, Ming-Ming Cheng, and Bo Ren. 2020. VecRoad: Point-based Iterative Graph Exploration for Road Graphs Extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 8910–8918.
- Vakalopoulou et al. (2015) Maria Vakalopoulou, Konstantinos Karantzalos, Nikos Komodakis, and Nikos Paragios. 2015. Building Detection in Very High Resolution Multispectral Data with Deep Learning Features. In IGARSS.
- Van Etten et al. (2018) Adam Van Etten, Dave Lindenbaum, and Todd M Bacastow. 2018. SpaceNet: A Remote Sensing Dataset and Challenge Series. arXiv preprint arXiv:1807.01232 (2018).
- Wang et al. (2013) Yin Wang, Xuemei Liu, Hong Wei, George Forman, Chao Chen, and Yanmin Zhu. 2013. CrowdAtlas: Self-Updating Maps for Cloud and Personal Use. In Proceeding of the International Conference on Mobile Systems, Applications, and Services (MobiSys). 27–40.
- Yang et al. (2019) Xiaofei Yang, Xutao Li, Yunming Ye, Raymond YK Lau, Xiaofeng Zhang, and Xiaohui Huang. 2019. Road Detection and Centerline Extraction via Deep Recurrent Convolutional Neural Network U-Net. IEEE Transactions on Geoscience and Remote Sensing 57, 9 (2019), 7209–7220.
- Zhou et al. (2018) Lichen Zhou, Chuang Zhang, and Ming Wu. 2018. D-LinkNet: LinkNet with Pretrained Encoder and Dilated Convolution for High Resolution Satellite Imagery Road Extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 182–186.