1,858
Views
1
CrossRef citations to date
0
Altmetric
Research Article

An investigation of the temporality of OpenStreetMap data contribution activities

, &
Pages 259-275 | Received 22 Jun 2021, Accepted 08 Sep 2022, Published online: 07 Oct 2022

ABSTRACT

OpenStreetMap (OSM) is a dataset in constant change and this dynamic needs to be better understood. Based on 12-year time series of seven OSM data contribution activities extracted from 20 large cities worldwide, we investigate the temporal dynamic of OSM data production, more specifically, the auto- and cross-correlation, temporal trend, and annual seasonality of these activities. Furthermore, we evaluate and compare nine different temporal regression methods for forecasting such activities in horizons of 1–4 weeks. Several insights could be obtained from our analyses, including that the contribution activities tend to grown linearly in a moderate intra-annual cycle. Also, the performance of the temporal forecasting methods shows that they yield in general more accurate estimations of future contribution activities than a baseline metric, i.e. the arithmetic average of recent previous observations. In particular, the well-known ARIMA and the exponentially weighted moving average methods have shown the best performances.

1. Introduction

For well more than a decade, OpenStreetMap (OSM) has been the world’s most successful volunteered geographic information project. Founded in 2004, the geographical coverage, geometric accuracy, semantic completeness, and overall reliability of the data have increased immensely since then (Arsanjani et al. Citation2015). Consequently, OSM has been a fundamental dataset in numerous geospatial research and applications, like pedestrian and motorized routing (Neis and Zipf Citation2008; Novack, Wang, and Zipf Citation2018), urban change modeling (Zhang and Pfoser Citation2019), land use and cover monitoring (Schultz et al. Citation2017), and fine-resolution population mapping (Bakillah et al. Citation2014), to name a few. However, geospatial research and applications based on OSM data frequently assume and rely on the stability of the data, neglecting the fact that it is constantly evolving and changing (Sieber and Haklay Citation2015). Even in large cities in the Global North, where data has been contributed and edited by larger numbers of contributors since longer date, OSM is a dataset in constant change.

One of the reasons for this is that the world itself is dynamic. Venues change their names, buildings are erected and extended, urban green areas are implemented or re-designed, etc. Also, OSM enables an always changing number of active volunteers to tag (i.e. to semantically annotate) features without any constraints. Although tag contributions and editions are usually intended to enhance the semantic completeness of map features (Hecht, Kunze, and Hahmann Citation2013) and to achieve tagging consistency across geographically, culturally, and politically related areas (Davidovic et al. Citation2016; CitationZia and Cakir), because volunteer contributors often have different worldviews and are motivated by different agendas (Anderson, Sarkar, and Palen Citation2019; Grinberger et al. Citation2021; Gardner et al. Citation2019), disagreements and negotiations may occur on which tags should be used or become deprecated (Mocnik, Mobasheri, and Klonner Citation2018), leading to map features having their tags edited, deleted, and extended. Other more exceptional reasons for OSM features being created and edited are disruptive (intentionally or not) mapping (Ballatore Citation2014; Juhász et al. Citation2020) and the aim of contributors to assist disaster managers in areas affected by, for example, earthquakes and floods (Ahmouda, Hochmair, and Cvetojevic Citation2018; Zook et al. Citation2010).

Because OSM data will continue to expand and to be essential for an increasing number and variety of geospatial applications, it is important to better understand its dynamic processes of change and evolution. This will enable researchers and practitioners behind OSM-based applications and services to assess the instability of the data, plan maintenance and updates, and design improvements to their tools and services more effectively. Accordingly, the motivations of and the research questions (RQs) addressed in this work are the following:

RQ 1 -

Better understand the temporal dynamic of OSM data contribution activities. What is the most consistent time interval for understanding such activities? In other words, how far in the past previous activities seem to affect present ones? How strong are those effects?

RQ 2 -

Uncover temporal correlations (i.e. apparent influences) between OSM data contribution activities. For example, what is the strength of correlation between the amount of new volunteers contributing data within a certain time span and the amount of tag editions occurring in a certain time span after that?

RQ 3 -

Investigate the extent to which we can forecast contribution activities in OSM. What is the prediction accuracy achievable with statistical and machine learning forecasting methods? How far in the future can we make such predictions? How the different prediction models compare in terms of prediction accuracy for different contribution activities?

For addressing these questions, we explore different time series statistics and forecasting methods. The remainder of this paper is structured as follows. The next section provides an overview of previous works related to temporal analysis and evolution of OSM data. Section 2 describes our experiment set up, namely, the data, considered variables, and test-sites. Section 3 describes the methods explored in this work. Section 4 presents the results obtained and Section 5 concludes the paper with a summary and outlook.

1.1. Related work

Due to the crowdsourced aspect of OSM data, numerous works have proposed practical and theoretical approaches to assess its quality (Sehra, Singh, and Rai Citation2017). A multidimensional notion, quality may refer, for example, to semantic veracity and completeness, as well as to geometrical consistency and accuracy, or to the usefulness (i.e. fitness) of the data to a specific purpose, such as pedestrian routing (Novack, Wang, and Zipf Citation2018). Generally, OSM data quality is assessed extrinsically, based on comparisons with authoritative and trustworthy data, or intrinsically, based on different characteristics and dynamics of the data itself (Degrossi et al. Citation2018; Fogliaroni, D’Antonio, and Clementini Citation2018; Mocnik, Mobasheri, and Klonner Citation2018; Senaratne et al. Citation2017). Intrinsic quality measures are typically applied when no authoritative data is available (or is comparable) or when the intention is to assess the fitness of the data to specific purposes. Substantial research efforts have focused on analyzing the temporal dynamics and historical evolution of OSM data as a way to intrinsically infer and assess its fitness and reliability. Raifer et al. (Citation2019), for example, developed the robust and flexible OpenStreetMap History Database (OSHDB) tool and application programming interface, which can support intrinsic data quality analyses based on the historical evolution of the data within user-defined polygons or administrative areas. The authors also provide a thorough overview of the characteristics and functionalities of other similar tools, such as Tag Info and OSM Tag History.

Other works have proposed quantitative ways of assessing the maturity and stability (reflecting completeness) of certain OSM features (e.g. roads, hospitals) based on the temporal cumulative function of a certain attribute of the data (e.g. the total length of roads in a given region). Barrington-Leigh and Millard-Ball (Citation2017) approximated this cumulative function with a sigmoid shape and detected saturation (representing stability) in ways emerging from the model, i.e. from the asymptote of the sigmoid. Previously, Grinberger et al. (Citation2021) had detected saturation simply through empirically defined saturation threshold. These and other related works have conducted analyses on OSM historical data not for understanding the dynamics, processes, and patterns through which the data evolves, but rather to infer the trustworthiness and completeness of the data.

The maturity or completeness of OSM data has been analyzed not only temporally but also spatially. Maguire and Tomko (Citation2017), for example, proposed metrics based on the temporal trends of the geometric definition of features to assess the spatially variable maturity of the data. Their approach is able to identify areas where mapping communities have reached consensus in the geometric definition of certain features as well as areas in need of more intense editorial activity in order to enhance the homogeneity of the data. Also aiming to better understand the spatio-temporal dynamics of the data contribution evolution of OSM, Arsanjani et al. (Citation2015) developed a cellular automata model. The model could also be used to predict potential future states of the data. So far though, the authors tested their model only in the medium-size city of Heidelberg (Germany). Furthermore, no validation of the results has been provided for their one-year ahead predictions. In a related paper, Arsanjani et al. (Citation2015) proposed a contribution index encompassing metrics of node quantity, interactivity, semantic completeness, and the ability to attract contributions. Besides analyzing the spatio-temporal evolution of the index for grid cells in Stuttgart (Germany), they also applied a cellular automata model to project the state of OSM data for that city up to the year of 2020. CitationZia and Cakir studied the spatial distribution of OSM road features contributions across time within Turkey. Similarly to the work of Corcoran, Mooney, and Bertolotto (Citation2013), they reported the two-phase evolution tendency of exploration, where features are discovered, followed by a densification phase, where features get connected to each other and their semantic information are deepened.

Other works have focused on the temporal evolution of the semantic information of OSM features, which is formalized as so-called tags. Mocnik, Zipf, and Raifer (Citation2017), for example, studied the temporal dynamics with which tags emerge, get consolidated, and finally documented in the OSM Wiki webpage, the OSM folksonomy portal serving as a catalog of generally accepted and suggested tags. By analyzing the temporal chart on the number of documented keys and tags since 2008 and fitting an exponential function to it, the authors also projected near future scenarios on the number of new documented tag keys. Along these lines, a number of tools have been developed and made available online that enable users to obtain temporal charts on the use of specific tags in OSM worldwide or for a given geographic area. These tools include TagInfo, OSM Tag History, and the Ohsome Dashboard.

Lastly, part of the OSM research community has focused on spatio-temporal patterns of OSM data contribution in areas affected by natural disasters, and where data had been typically scarce previous to the calamity (Ahmouda, Hochmair, and Cvetojevic Citation2018; Xu, Li, and Zhou Citation2017). Xu, Li, and Zhou (Citation2017) investigated the temporal and spatial correlation between the creation of road and building features after the 2016 hurricane Matthew in Haiti. Ahmouda, Hochmair, and Cvetojevic (Citation2018) focused on short- and long-term changes in mapping community composition and behavioral due to earthquake events in Nepal and Italy.

2. Preliminaries

2.1. Data acquisition and sampling

The aim of this paper is to propose simple quantitative ways of investigating the temporality of OSM data contribution activities. More specifically, we focus on contribution and edition activities of points-of-interests (POIs), which we define as OSM nodes containing tag keys “name” and “amenity”. Readers unfamiliar with the concept of OSM tags and with OSM’s feature types are referred to its Wiki page (Novack et al., Citation2017).

For conducting our experiments, the complete POI dataset at January 2020 from 20 cities worldwide was collected. The aspects considered for selecting these cities, shown in , were their cultural and socio-economic diversity, their national importance, and their well-established community of active OSM mappers. Next, the editing history of each POI in each of the 20 cities was extracted. This historical record describes how the POI has changed over time since its creation, which we denote as its first entry. All subsequent entries are modifications and/or addition of data, that is, tag deletions, tag insertions or tag editions, as well as modifications of the POI’s geographical location. (A POI can also be completely deleted at some point in time. However, since only existing POIs by January 2020 were considered in our experiments, only POIs creations and modifications were detected and considered in the construction of the POIs’ historical records.) More formally, a historical record is comprised of a set of entries {e1,e2,,en}, where e1 represents the creation of the POI. Each entry has an associated timestamp. Extracting the edition variables we consider simply required comparing the states of a POI at entries ei and ei1, extracting the timestamp of ei, and associating the change of state from ei1 to ei to its respective indexed month, as discussed below. At each point in time, a POI can be created or modified. Although in reality, both actions co-occur at a POI’s e1, an exclusive distinction was made for the convenience of modeling. Therefore, an entry can refer to either a POI creation or modification. In the latter case, an entry can entail more than one type of edition. gives the description of the variables computed at each entry ei of each POI from each of the 20 cities we consider in this study. Thus, an entry is a vector of seven binary elements, each referring to one of the variables in . From now on, we refer to the variables through their alias, also presented in .

Table 1. The 20 cities whose entire POI dataset at January 2020 and its editing history were considered in this study.

Table 2. Variables computed at each entry from each POI. The variables were computed by comparing two subsequent states of the POIs’ historical record.

Following, the time series of each variable and city was derived from the entry vectors of all POIs of that city. The time series were sampled, i.e. discretized, into time slots of one month. The value of a variable for a specific month was computed by simply summing all respective elements from all entry vectors whose timestamps are within that month. In this way, a time series was created for each city and each of the variables presented in . All time series originally spanned from January 2007 to January 2020. However, for modeling purposes, the year of 2007 was omitted from all time series, as they contain very little data points for most of the cities. Also, time slots with values above the 99% quantile for any of the seven variables were considered outliers and had theirs values substituted by the value of its immediately preceding slot. In OSM, these outliers are caused mainly by mapping parties or major data imports (Juhász and Hochmair Citation2018), which we regard as exceptional events.

For the sake of reproducibility, and in accordance with the spirit of open science, all data processing procedures described in this paper are reproducible with a library coded in Python and made openly available at https://github.com/le0x99/OSMEvolution. This Python library works just as well with other OSM geometries, namely, ways and relations, and for any given city or OSM-mapped geographical boundary. At the ”How it works” section of the repository, the reader will also find a detailed technical description of how the OSM data was queried and structured into the time-series analyzed in this study. A description of this workflow is also given in .

Figure 1. Data querying and structuring workflow for generating the time-series used in this study.

Figure 1. Data querying and structuring workflow for generating the time-series used in this study.

3. Methods

In this section, we introduce the methods used for addressing the RQs presented at the introduction of this paper. The methods applied to tackle RQ 1 are described in Sections 3.1 and 3.2. RQ 2 was approached with the method described in Section 3.3. Lastly, the methods addressing RQ 3 are described in Sections 3.4 and 3.5.

3.1. Rolling window

For addressing RQ 1, related to the most consistent time interval with which OSM data contribution activities take place, the simple rolling window method was applied. Let yti be an observation y of variable i at time slot t. As explained above, for each time slot and variable, the element referring to that variable in all entry vectors whose timestamp is within that slot are summed to yield the observation y. The rolling window method estimates this observation as yˆti=1Wl=1Wytl, in words, as the average of the observations of i at previous time slots t1,t2,tW, where W denotes the size of the window. The squared error metric (ytiyˆti)2 can be used to measure the difference between an observation at time t and the average of its W previous observations. After computing the squared error for all observations of a time series, we can compute its mean square error (MSE). The MSE metric was computed for all seven variables, all 20 cities, and for different W window sizes. Given a time series of a variable i and a certain city, the W for which the lowest MSE is obtained is the most consistent time span with which that variable changes in time. That gives us valuable insights into that variable’s temporal dynamic. For example, if W=4 yields the lowest MSE, we can conclude that the data generating process of variable i is, in average, most correlated with and influenced by activities occurring in the last four time slots, whether these are days, weeks, or months.

3.2. Time Series Decomposition

Also related to RQ 1’s motivation to better understand the temporal dynamic of OSM data contribution activities, time series decomposition was applied. As the name implies, this is a procedure for decomposing a time series into different components, namely, trend, seasonality, and residuals. The first two components relate to two underlying properties of the data we aim to analyze. These are, respectively, the strength of increase in data contribution and edition activities in OSM, and how cyclical these activities are.

In this work, an additive decomposition technique was adopted, such that any observation yti can be recomposed through the sum of its three components. Formally,

(1) yti=Tti+Sti+Rti,(1)

where Tti represents the trend component, Sti the seasonal component, and Rti the residuals. In order to compute the components, the cycle length m must be defined. In our case, for monthly aggregated data and an assumed yearly cycle, we defined m=12. The trend component Tti was computed as the 2m (i.e. 24-month) moving average of the data. For estimation of the seasonal component, the de-trended time series ytiTti was averaged for every month in the cycle, meaning that, for a given month, we obtain all values of the de-trended time series that lie within that month and take their average. The outcome of this are 12 values, which are repeated for every year and express the time series seasonality component. The third component, Rti, is the residual between the underlying original time series and the trend and seasonal components, such that Rti=yti(Tti+Sti). This component expresses all remaining properties of the time series, but it can also be viewed as the deviation from a simple model, based on which each time series is only characterized by trend and seasonality in an additive way.

Following the approach proposed by Wang, Smith, and Hyndman (Citation2006), the strength of the trend FT and seasonality FS components of the variables’ time series can be quantified, respectively, according to the equations below:

(2) FT=max0,1VarRtVarTt+Rt(2)

and

(3) FS=max0,1VarRtVarSt+Rt,(3)

where, as before, Tt, St, and Rt stand for the trend, seasonality, and residuals components extracted as discussed above. In time series with strong trend, the variance of the trend component is significantly larger than the variance of the residuals; therefore, FT will be close to 1. Likewise, in time series with strong seasonality, the variance of the de-trended data is larger than the variance of the residuals, leading to FS closer to 1.

3.3. Temporal linear correlation

For uncovering temporal correlations between OSM data contribution activities, (which is the rationale of RQ 2), the Pearson’s correlation coefficient ρ was used. Besides measuring the linear correlation between two different data contribution variables, ρ was also used for measuring the correlation between observations yti of a variable i at time t and their temporally lagged values ytTi. Formally, this autocorrelation is given by

(4) ρ(yti,ytTi)=cov(yti,ytTi)σytiσytTi,(4)

where T is the temporal lag, the nominator is the covariance between yti and ytTi, and the denominator is the product of their standard deviations.

This simple metric enables inspecting, for different T>0, how temporally auto-correlated a variable i is. It also enables, and was used for, as mentioned, investigating the temporal correlations between different data contribution variables. For example, we can measure the correlation of tag insertions at t1 and tag editions at t. Note that we can consider in this analysis different T>=0 for all pairs of i and j variables.

3.4. Temporal Forecasting Methods

To address RQ 3, the temporal forecasting methods presented in Sections 3.4.2 to 3.4.6 below were applied. In the next section, the notation used for presenting these methods and the metric to evaluate their performance is introduced.

3.4.1. Adopted notation

For each of the 20 cities considered in this study, four multivariate time series were generated, each for all seven variables presented in and discretized into time slots of 1–4 weeks. Thus, given city c and time sampling h (in 1–4 weeks), we denote the respective time series as Xc,h. The series contains T observations XtXc,h, which are row vectors containing the values of the seven variables for time slot T. Note that the size of T depends on h. This vector is referred to as the objective vector, as it holds the seven target variables we aim to forecast. A specific target variable will be denoted as xt,iXt, where xt,i is the column vector of the specific target variable i with length T. Each of the methods explored and compared regarding their performance in forecasting the objective vector Xt will be introduced in the following sections. If the method iteratively approximates each variable xt,i separately, instead of the joint Xt vector, the target variable xt,iXt is then notated as Xt.

3.4.2. Moving average methods

The first and most simple time series forecasting method explored in this study is referred to as the naive method (NM). It approximates the unobserved value of variable x at time t as its previous observation xt1. Thus, this method can be formalized as follows:

(5) NMxtxt1.(5)

Despite its simplicity, the NM is often used in time series forecasting and is in fact hard to be outperformed by more complex methods. This is due to the often existing temporal correlation of time series observations.

The next three forecasting methods we present in this section can be generally regarded as moving average methods. The first of these is the simple moving average (SMA) method, which is an extension of the naive baseline method introduced above. It estimates an observation xt as the arithmetic mean of its previous xt1,xt2,,xtW observations, where W is the size of the lagged window. The linearly weighted moving average (LWMA) method is a derivation of the SMA method in that it assigns to the observations within the lagged time window W linearly decreasing weights, such that more recent observations receive higher weights than less recent ones. Given a window W and lag l[0,W], the linear weight for observation xtl is defined as

(6) wl=2(Wl+1)W(W+1).(6)

Alternatively to a linear decay of the weights wl, these can also decay exponentially. The exponentially weighted moving average (EWMA) method implements this idea. Thus, the exponential weight for a lag l can be approximated by

(7) wl=(1α)1Wαll=1W(1α)1Wαl,(7)

where α controls the initial weight of the first lag.

3.4.3. Ordinary Least Squares Autoregression

A widely used method in univariate time series forecasting is the ordinary least square autoregressive model (AR). It estimates an unobserved objective vector xt as a linear function of its own lagged values xt1,xt2,,xtl, where l denotes the maximum incorporated lag. The assumption of the AR model is that a time series variable can be estimated as a linear function of its l lagged values. Formally, the AR model is a linear regression one and has the following form:

(8) ARxtw0+l=1Lwlxtl+\isint,(8)

where \isintN(0,σ2) is irreducible white noise and wl is the ordinary least square coefficient of the lth lagged observation.

Before the AR can be correctly applied, it is necessary to address the assumed stationarity of the underlying data. Stationarity refers to when the mean and variance of the data are constant over time and it is assumed by all models within the AR-family. As expected from crowdsourced data production processes, however, a certain seasonality (non-uniform variance over time) and trend (non-uniform mean over time) can be observed in our time series. In order to address this, the first-order difference Δxt=xtxt1 was applied to our data. With that, EquationEquation 8 was altered to

(9) ARΔxtw0+l=1LwlΔxtl+\isint,(9)

which is trivial to re-transform for later prediction. Please consider that the first-order difference is also applied to the other models from the AR-family presented below.

3.4.4. Autoregressive Moving Average

The Autoregressive Integrated Moving Average Model (ARIMA) is an extension of the AR model above and is extensively used in the modeling of time-varying processes. Like the AR model, ARIMA entails an autoregressive part. Additionally to the AR part though, the ARIMA model includes a moving average (MA) term that models xt as a linear function of the current and previous white noise error terms \isint,\isint1,,\isintl, where l parameterizes the model, such that the model MA(l) is formalized as xtμ+εt+β1εt1++βlεtl, where μ is the mean of the series. In contrast to the AR part of the model, fitting the MA term cannot rely simply on the Ordinary Least Square measure, since the lagged white error terms are not directly observable. The error terms are therefore fitted iteratively using Maximum Likelihood Estimation. Combining the AR and MA terms, before applying first-order differencing, leads to the formalization of the ARIMA model as:

(10) ARIMAΔxtμ+εt+l=1pαlΔxtl+l=1qβlεtl(10)

3.4.5. Vector Autoregression

Similarly to the AR and ARIMA models, the Vector Autoregression (VAR) model is a multivariate statistical time series forecasting method based on the idea that a set of variables xtXt can be estimated as a function of its lagged values Xt=f(Xt1,Xt2,,Xtl), where l represents the maximum lag in time. In contrast to both AR and ARIMA though, VAR estimates a variable xtXt not only as a linear combination of its lagged values but as a linear combination of all other lagged variables xtXt. Technically, the VAR model solves a set of linear equations, each of which being an Ordinary Least Square linear regression with multiple inputs (i.e. the lagged values of Xt are the input and Xt is the output). The VAR model is formulated as

(11) VARΔXtw0+w1ΔXt1+w2ΔXt2++wLΔXtL+\isint,(11)

where each wl is a 7×7 weight matrix (7 is the size of our objective vector) and \isint is irreducible white noise, which in case of VAR is a vector with seven elements.

3.4.6. Random Forest

The last model we experimented with in this paper is the Random Forest (RF) method (Breiman Citation2001). Known for its efficacy in modeling highly non-linear relations in the data, RF is a method based on an ensemble of bootstrapped fully grown regression trees. Each split at each node is defined based on the criterion of the mean squared error. As recommended by Breiman (Citation2001), the number of features considered at each tree’s split was of Nx3, where Nx is the total number of input features. For prediction, the mean value of the set of fully grown trees is then used. A deeper understanding on the underlying tree-building algorithm can be obtained at Breiman’s seminal work (Citation1984).

Although not commonly applied for time series data forecasting, we decided to experiment with the RF regression method because it works fundamentally different than the statistical methods introduced above and may be effective in forecasting non-linear time-varying processes. In contrast to VAR and ARIMA, RF does not assume any distributional properties of the underlying data other than representative sampling.

In order to improve the forecasting ability of the RF model, temporal indicator features were added to the input feature vectors. Specifically, the following modal features were added: week of the year, in range [1,52], and day of the month, in range [1,31]. Also, the non-modal feature month of year was added, yielding a dummy variable, i.e. in range 0,1, for each of the 12 months. Thus, we are assuming that time-specific properties of the observations, like the month or the week, contain information that would be otherwise lost without these additional features. We experimented with two kinds of RF models, namely, the Univariate Random Forest (URF), which considers only the past values of the variable it predicts, and the Vector Random Forest (VRF), which considers the entire vector of past observations. The motivation for applying and comparing these two models is that, similarly, the AR and VAR models also consider, respectively, uni- and multivariate lagged observations, assuming, however, linearity of the data, differently from the RF models, as noted.

3.5. Forecasting models training, evaluation, and test

This section presents the procedure and metrics adopted to evaluate the performance of the forecasting methods presented above and implemented to address RQ 3.

3.5.1. Data partitioning

The time series generated as presented in Section 2 were divided into three parts. The first part, spanning from January 2008 to January 2018, was used for the training of the models. The second part of the data, spanning from January 2018 to January 2019, was used for evaluating the parameters of the models, as will be discussed below. After fitting the parameters, based on the criterion of performance maximization, the third part of the series from January 2019 to January 2020 was used for testing the performance of the models. The total number of observations of each of the data parts is the same for all the 20 cities, but depending obviously on the time sampling and forecasting horizon, which we henceforth note as h. shows the number of observations per city of each of the data partitions and time sampling rate forecasting horizons (in weeks). Note that, for example, when h = 1, there are 524 observations for training and 52 for testing for each of the 20 cities.

Table 3. Number of observations per city of the three data partitions and for the four time sampling rates and forecasting horizons (in weeks).

By distinguishing between evaluation and testing data, hard-coded parameters are avoided, as it is always possible to find a set of parameters that maximize any metric when the data they are evaluated on stays fixed. Although in this particular scenario, this may not be the case, even without an additional testing set, since we already cross-validate over the cities, we consider that this partition is still useful, as it further increases the reliability of the obtained results.

3.5.2. Model parameters evaluation

For each of the methods presented above, parameters were set based on a parameter evaluation analysis. The statistical models AR, VAR, and ARIMA require setting the maximum lag L, which essentially defines the model. For that, the evaluation dataset from each of the 20 cities was used to compute the average error and its standard deviation for each considered value of L. This procedure was undertaken independently for all four time sampling rates of 1–4 weeks. Likewise, concerning the ARIMA model, the value of L was defined based on the evaluation dataset. Regarding the methods based on rolling statistics, i.e. SMA, LWMA, and EWMA, the value of w, that is, the lagged window from which the moving average is calculated, was set. Concerning the EWMA model, the α parameter was set.

3.5.3. Method performance metric

Different performance metrics have been proposed in the literature of time series forecasting, each having its pros and cons (Armstrong and Collopy Citation1992; Makridakis Citation1993; Makridakis and Hibon Citation1979). The metric used in this work is based on the squared error of the predicted objective vector, given city c and time sampling h. Given the true vector Xt and the predictions Xˆt, this metric is defined as:

(12) \isintc,h=(Xtc,hXˆtc,h)2,(12)

where \isin represents the squared error vector containing the squared errors of all seven objective variables within Xt. Based on this metric, the conditional RMSE vector, meaning the RMSE given time sampling h and city c, was computed. The conditional RMSE is formalized as:

(13) RMSEc,h=1Tht=1Th\isintc,h,(13)

where Th denotes the total amount of test observations given h (e.g. Th=1week=52 observations per city). The RMSE was used instead of the plain MSE in order to get interpretable results when analyzing the errors. Since RMSEc,hR7, we are able to inspect each objective vector element by element. By using this conditional error metric, we can compare the RMSE vectors of all target variables within Xt conditioned on city c and horizon h. For any model M, we now can inspect RMSEMc,h, the models main performance criterion. However, using the conditional RMSE in a comparative analysis is not trivial, as we observe different value ranges for (i) the different cities (e.g. Berlin has a larger value range than Heidelberg), (ii) the chosen forecasting horizon h (monthly indexed data has obviously another range than the weekly one), and, most importantly, (iii) the objective variable themselves (e.g. tag additions has a higher value range than locational change, for example). Now, comparing the conditional performance of the models is hampered by the arbitrary scaling of the RMSE, which prevents a meaningful comparison across the different objective variables. Although there exists a number of methods to normalize the RMSE, all of them have drawbacks. In this work, we chose to simply divide a model’s RMSE by the RMSE of the naive baseline model NM with respect to horizon h, city c, and the objective variable i. With that, a formally consistent and easily interpretable error metric is obtained. Thus, given a model M and its conditional error vector RMSEMc,h, the normalized error vector E for any model is given by

(14) E(M,c,h)=RMSEMc,hRMSENMc,h,(14)

such that

(15) E(NM,c,h)=[1,,1]c,h.(15)

Accordingly, Ei <1 indicates for any of the models a better performance than the baseline model NM and Ei >1 a worse performance concerning variable i.

4. Results

In this section, we present the results obtained in our analyses. Subsections 4.1, 4.2, and 4.3 relate, respectively, to RQs 1, 2, and 3. Reflections on the findings to be presented and their utility for the OSM community and practitioners are discussed in Section 5.

4.1. The temporal dynamic of OSM data contribution activities

Referring to the rolling window method presented in Section 3.1 and related to RQ 1, shows for different window sizes (in months) and for all seven variables the mean and 95% confidence interval of the RMSE metric computed from the time series of the 20 cities considered in this study. In order to make the results comparable, as the variables have different ranges across cities of different sizes and across themselves, the values plotted in were all normalized to the mean RMSE, across all 20 cities, of each variable for the window size of 2 months. Aiming a better interpretation and a more structured discussion of the results, these variables were grouped into activity and modification variables and shown separately in . We maintain this grouping for most of the discussions below.

Figure 2. For all seven variables, regarding the 20 cities considered in this study, the figure shows the normalized mean and 95% confidence interval of the rolling window metric RMSE (Section 3.1).

Figure 2. For all seven variables, regarding the 20 cities considered in this study, the figure shows the normalized mean and 95% confidence interval of the rolling window metric RMSE (Section 3.1).

The window size for which the normalized mean RMSE is the lowest can be interpreted as the most consistent length of time each variable’s evolution, i.e. changes in time, can be tracked. In the formulations of RQ 1: as the most consistent time interval for understanding OSM mapping activities and as how far in the past such previous activities seem to affect present ones. One can observe from that the minimum normalized mean RMSE for all variables lies between 4 and 6 months, with the exception of variable tag deletion (around 13 months). In other words, these activities happening in any given month correlate most strongly with such activities happening, in average, in the previous 4–6 months. This uncovers the temporal ”rhythm” of these OSM data contribution activities. That tag deletion has a different temporality, i.e. a slower pace of change, is not surprising. Its minimum normalized mean RMSE of approximately 13 months indicates a reference of time for tags to become deprecated and for mappers focusing on data correction to become more active. It is also noticeable, and expected, that the variance of the normalized mean RMSE, as indicated by the 95% confidence intervals, increases consistently with larger window sizes.

Referring to the time series decomposition method presented in Section 3.2, shows the mean and 95% confidence interval (across the 20 cities) of the decomposed trend of each variable for each month of the time series. For computing the decomposed trends, a moving average window size of 12 months was considered. The decomposed trend values on the y-axis of are standardized according to the equation below:

(16) stand.(yti)=ytiμyiσyi,(16)

Figure 3. Mean and 95% interval of the trend component of the time series of the 20 cities for all seven variables. The trend component was computed with a 12-month average window. The values on the y-axis are standardized according to EquationEquation 12.

Figure 3. Mean and 95% interval of the trend component of the time series of the 20 cities for all seven variables. The trend component was computed with a 12-month average window. The values on the y-axis are standardized according to EquationEquation 12.(12) \isintc,h=(Xtc,h−Xˆtc,h)2,(12)

where, as before, yti is an observation of variable i at time slot t. μ and σ are, respectively, the mean and standard deviation of all observations from the time series. Besides enabling the graphical presentation of all variables together, the normalization allows for statistical aggregation of the decomposed trend of cities of different sizes as well as the comparison of the decomposed trend of variables with very different value ranges. It can be seen, with the exception of variable new mapper which exhibits an outstanding pick on year 2017, that all variables exhibit a similar mildly linearly increasing trend with constant confidence intervals. The decomposed trend lines serve as evidence that a linear model is adequate for projecting OSM data contribution activities for future years.

presents the descriptive statistics of the R-squared indexes computed by fitting, for each variable and city, an ordinary least squares regression model in which the independent variable is the time-indexed months of the series and the dependent variable is the respective decomposed trend of the contribution variables. The R-squared of such models is a reliable metric of the strength of linearity of the trends. It can be seen from that the mean of the R-squared values obtained from the 20 cities is for all seven variables high (above 0.58 for four out of seven variables). The maximum R-squared values are all above 0.80 and variable new mapper exhibits a particularly strong linear trend, with an average R-squared value above 0.70 for the 20 cities. That further supports, with quantitative evidence, the argument that OSM data contribution activities evolve, in general, linearly, thus shedding light on the nature of temporal dynamic of OSM data contribution activities.

Table 4. Descriptive statistics of the R-squared values obtained by linearly regressing, for each variable and city, the time-indexed months of the series with their respective decomposed trend values.

We also analyzed the intra-annual dynamic of each variable based on the time series’ seasonality component, extracted with the time decomposition method presented in Section 3.2. shows, for each of the seven variables, the box plot distributions, across the 20 cities, of the mean seasonality of each month of the year. As before, to account for the different variable ranges and city sizes, the seasonality values are standardized according to EquationEquation 16. It can be seen that the activity variables have a stronger intra-annual contribution cycle, similar to a sine wave. Such cyclical pattern is less prominent among the modification variables.

Figure 4. Box plot distributions of the mean seasonality component for each month and variable across the 20 cities. The seasonality component values are standardized, according to EquationEquation 12, for allowing comparison.

Figure 4. Box plot distributions of the mean seasonality component for each month and variable across the 20 cities. The seasonality component values are standardized, according to EquationEquation 12(12) \isintc,h=(Xtc,h−Xˆtc,h)2,(12) , for allowing comparison.

shows the box plot distributions, across the 20 cities, of these components’ strength for all seven variables. In line with the results from , the trend strength for all variables, with the exception of tag deletion, is high, specially for the variable new mapper. The seasonality strength, on the other hand, is comparatively lower for all seven variables. That variables create and positional change exhibit the strongest seasonality is also reflected in , where, for these two variables, a stronger intra-annual variance across the monthly distributions of the standardized seasonality can be observed.

Figure 5. Box plot distributions, across the 20 cities, of the trend and seasonality strength for all seven variables.

Figure 5. Box plot distributions, across the 20 cities, of the trend and seasonality strength for all seven variables.

The results presented in this section uncover interesting temporal patterns of OSM data evolution. In summary, for most variables considered, changes occur in a pace of 4–6 months and in linear way. Although some strength of seasonality is observed, the linear trend, throughout more than 10 years, is what stands out. That constant linear increase identified in our data contests the previously adopted S-shaped function to model the growth of OSM data (Gröchenig, Brunauer, and Rehrl Citation2014, Citation2014).

4.2. Temporal correlations among OSM data contribution variables

In order to uncover temporal correlations between OSM data contribution activities (RQ 2), cross auto-correlations (see Section 3.3) were computed between all possible pairs of activity and modification variables in time lags of 0–8 months. The rationale of this analysis is to provide evidence on possible causal relations between OSM mapping activities. shows the box plot distributions of these correlations between the activity variables across the 20 cities considered in this study. shows the same for the modification variables.

Figure 6. Box plot distributions, for all 20 cities, of the autocorrelation between observations from all pairs of activity variables in time lags of 0–8 months.

Figure 6. Box plot distributions, for all 20 cities, of the autocorrelation between observations from all pairs of activity variables in time lags of 0–8 months.

Figure 7. Box plot distributions, for all 20 cities, of the autocorrelation between observations from all pairs of modification variables in time lags of 0–8 months.

Figure 7. Box plot distributions, for all 20 cities, of the autocorrelation between observations from all pairs of modification variables in time lags of 0–8 months.

In , one sees that new mapper is the variable with strongest autocorrelation. Also, and although subtle, it is noticeable that there is less variance in the correlations computed between modify and new mapper than in other pairs of so-called activity variables. That indicates a stronger and more stable relation of influence between new mappers and feature modification than feature creation activities, meaning that these new mappers initial activities tend to be to edit as opposed to create map features. In , it can be seen that tag insertion is the variable with strongest autocorrelation and that a stronger mutual influence is observable between variables tag insertion and tag edition. This indicates that mappers inserting tags might be also editing tags at the same and possibly other features. In both figures, as one would expect, the correlations tend to decrease, and variances tend to increase, with longer time gaps.

4.3. Forecasting OSM data contribution activities

Following the procedures introduced in Section 3.5, the forecasting methods presented in Section 3.4 were evaluated and compared. In , the forecasting performances obtained for each contribution variable, by each forecasting method, and for forecasting horizons h of 1–4 weeks are shown. The box plot distributions refer to the performances obtained across the 20 cities. The red line where the normalized error is 1 refers to the performance of the baseline method NM (see EquationEquation 3). The fact that the forecasting performance metric is relative to the baseline method enables the aggregation of cities of different sizes and magnitude of contribution activities. The models' absolute errors can be obtained by referring to in the Appendix, where the descriptive statistics of the absolute errors of the baseline method for all variables, cities, horizons, and samples considered in this study are shown.

Figure 8. Performance of the forecasting methods for the seven variables and four forecasting horizons (in weeks). The box plot distributions refer to the 20 cities and the red-dotted line refers to the naive baseline method.

Figure 8. Performance of the forecasting methods for the seven variables and four forecasting horizons (in weeks). The box plot distributions refer to the 20 cities and the red-dotted line refers to the naive baseline method.

It can be seen that as the forecasting horizons increase from 1 to 4 weeks, the average performance of each method for each variable decreases and the standard deviation thereof increases. In other words, predictions are less reliable the further ahead we attempt to forecast the activities. However, performances are in general better than the baseline model, leading us to the conclusion that forecasting methods give us a better prediction of contribution activities than the baseline approach. The ranking of the performance of these methods is very similar for the seven variables and for the four different forecasting horizons. Most often, the best performance method is the ARIMA, followed by EWMA and LWMA. Also noticeable is that variables “new mapper” and “tag deletion” are frequently better forecast than the other variables across the different models and horizons. Due to the differences and complexities of the models, it is hard to pinpoint the reasons for this, but they may have to do with the fact that “new mapper” is notably the variable with strongest autocorrelation () and “tag deletion” has the lowest mean normalized RMSE (). On the other hand, “create” and “tag insertion” are the variables for which the models are least accurate. This might be related with “tag insertion” having the highest mean normalized RMSE () and “tag deletion” the strongest seasonality () and variation thereof () across the cities.

Finally, shows the mean and standard deviation of the normalized error metric (EquationEquation 14) of all eight forecasting methods for each of the 20 cities and the four forecasting horizons considered in this study. It can be noticed that the trend of higher errors for longer horizons holds somewhat consistently for all cities. Importantly, the size of the city does not seem to affect the prediction accuracy of the models.

Figure 9. Mean (a) and standard deviation (b) of the normalized error metric of all eight forecasting methods for each of the 20 cities and the four forecasting horizons considered in this study.

Figure 9. Mean (a) and standard deviation (b) of the normalized error metric of all eight forecasting methods for each of the 20 cities and the four forecasting horizons considered in this study.

5. Summary and outlook

OSM is a constantly changing crowdsourced dataset representing a constantly changing world. It is maintained by a dynamic community of volunteers that define and redefine, not without negotiations and conflicts, how new and persisting features should be mapped and edited. In this work, we explored time series analysis methods and metrics to shed light on the temporal dynamics of different OSM data contribution activities. The trend of linear increase of these activities, their moderate but perceivable intra-annual cycle, and their temporal auto- and cross-correlations are identified patterns that help us better understand the nature of this valuable dataset and the community behind it. These patterns may be considered when planning maintenance and updates of geospatial applications and services based on OSM data, as they enable estimating how much the data will change in a given time (in terms of spatial coverage and semantic redefinition of geo-concepts). The OSM community of engaged mappers and moderators also benefit from evidence-based hints on when and to what type of activities it should pay attention to, indicated by the seasonality graphs and the identified temporal correlations between different data edition and contribution activities. Furthermore, the community can get valuable warnings on significant deviations from the edition patterns identified by the temporal trend and the rolling window graphs, thus enhancing its situation awareness and data quality control efficacy. Importantly, all the graphs generated and used metrics in this study can be reproduced for specific regions and for specific types of features, such as roads, buildings, parks, etc.

We also investigated the extent to which OSM data contribution activities can be forecast. Besides applying a variety of established time series regression methods to forecast seven different OSM data contribution activities in prediction horizons of 1–4 weeks, we proposed an interpretable metric to compare the performance of the models in forecasting these different activities. The Python code developed for extracting and structuring historic OSM data into time series, as well as for applying the compared forecasting methods, is as mentioned, readily accessible to anyone aiming to reproduce this forecasting analysis in other areas and on any set of OSM objects.

Although large data imports and data contribution parties can be expected to negatively affect the accuracy of the forecasting models and the reliability of temporal correlation metrics and patterns, previous studies have reported that such events have long-term influence on the mappers activities and on data contribution dynamics (Ahmouda, Hochmair, and Cvetojevic Citation2018; Gröchenig, Brunauer, and Rehrl Citation2014; Juhász and Hochmair Citation2018). Thus, these events do not hamper the applicability of the analyses and methods explored in this study.

While our work addresses the gaps in current OSM literature of forecasting contribution activities and investigating large-scale temporal data development patterns across cities from different cultural contexts, this effort is far from complete and this paper can be considered a first exploratory investigation. A particularly interesting follow-up study is to apply the approaches presented in this work to non-urban areas. Also, it is relevant to investigate how groups of cities of different cultural contexts, sizes, national importance, and levels of developed OSM community compare in terms of the patterns uncovered in our work. Furthermore, many possibilities are open in terms of considering other attributes of the OSM features. These include their spatial context, semantic categories, individual edition history, and characteristics of the mappers acting on them. Besides capturing important nuances of the data temporal development, these mapper- and feature-related attributes may also increase the forecasting accuracy of the models and extend their applicability to temporal horizons beyond 4 weeks. We acknowledge that predicting data contribution activities, somewhat reliably, for a horizon of a few weeks only calls for improvements. In this regard, besides these spatial, semantic, and historical attributes, arguably more powerful forecasting methods, although hardly interpretable, may yield longer-term and more accurate predictions. Specifically, Long Short-term Memory models have shown impressive performance in time series forecasting in different domains (Hua et al. Citation2018; Lindemann et al. Citation2021).

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement

The data that support the findings in this study, as well as the code to process it with the methods presented in this paper, are available in Github at https://github.com/le0x99/POI-Evolution-Forecasting.

References

  • Ahmouda, A., H. H. Hochmair, and S. Cvetojevic. 2018. “Analyzing the Effect of Earthquakes on OpenStreetMap Contribution Patterns and Tweeting Activities.” Geo-Spatial Information Science 21 (3): 195–212. doi:10.1080/10095020.2018.1498666.
  • Anderson, J., D. P. L. Sarkar, and L. Palen. 2019. “Corporate Editors in the Evolving Landscape of OpenStreetMap.” International Journal of Geo-Information 8: 232. doi:10.3390/ijgi8050232.
  • Armstrong, J., and F. Collopy. 1992. “Error Measures for Generalizing About Forecasting Methods: Empirical Comparisons.” International Journal of Forecasting 8 (1): 69–80. doi:10.1016/0169-2070(92)90008-W.
  • Arsanjani, J. J., M. Helbich, M. Bakillah, and L. Loos. 2015. “The Emergence and Evolution of OpenStreetMap: A Cellular Automata Approach.” International Journal of Digital Earth 8 (1): 76–90. doi:10.1080/17538947.2013.847125.
  • Arsanjani, J. J., P. Mooney, M. Helbich, and A. Zipf. 2015. “An Exploration of Future Patterns of the Contributions to OpenStreetMap and Development of a Contribution Index.” Transactions in GIS 19 6: 896–914. doi:10.1111/tgis.12139.
  • Bakillah, M., S. Liang, A. Mobasheri, J. J. Arsanjani, and A. Zipf. 2014. “Fine-Resolution Population Mapping Using OpenStreetMap Points-of-Interest.” International Journal of Geographical Information Science 28 (9): 1940–1963.
  • Ballatore, A. 2014. “Defacing the Map: Cartographic Vandalism in the Digital Commons.” The Cartographic Journal 51 (3): 214–224. doi:10.1179/1743277414Y.0000000085.
  • Barrington-Leigh, C., and A. Millard-Ball. 2017. “The World’s User-Generated Road Map is More Than 80.” PLOS ONE 12 (8): 1–20. doi:10.1371/journal.pone.0180698.
  • Breiman, L. 2001. “Random Forests.” Machine Learning 45 (1): 5–32. doi:10.1023/A:1010933404324.
  • Breiman, L., J. Friedman, C. J. S. Friedman, and R. A. Olshen. 1984. Classification and Regression Trees. 1st ed. Boca Raton, USA: Chapman Hall.
  • Corcoran, P., P. Mooney, and M. Bertolotto. 2013. “Analysing the Growth of OpenStreetMap Networks.” Spatial Statistics 3: 21–32. doi:10.1016/j.spasta.2013.01.002.
  • Davidovic, N., P. S. L. M. M. Mooney, L. Stoimenov, and M. Minghini. 2016. “Tagging in Volunteered Geographic Information: An Analysis of Tagging Practices for Cities and Urban Regions in OpenStreetMap.” ISPRS International Journal of Geo-Information 5: 232. doi:10.3390/ijgi5120232.
  • Degrossi, L. C., J. Porto de Albuquerque, R. D. Santos Rocha, and A. Zipf. 2018. “A Taxonomy of Quality Assessment Methods for Volunteered and Crowdsourced Geographic Information.” Transactions in GIS 22 2: 542–560. doi:10.1111/tgis.12329.
  • Fogliaroni, P., F. D’Antonio, and E. Clementini. 2018. “Data Trustworthiness and User Reputation as Indicators of VGI Quality.” Geo-Spatial Information Science 21 (3): 213–233. doi:10.1080/10095020.2018.1496556.
  • Gardner, Z., P. Mooney, S. DeSabbata, and L. Dowthwaite. 2019. “Quantifying Gendered Participation in Open-streetmap: Responding to Theories of Female (Under) Representation in Crowdsourced Mapping.” GeoJournal 85: 1603–1620.
  • Grinberger, A. Y., M. Schott, M. Raifer, and A. Zipf. 2021. “An Analysis of the Spatial and Temporal Distribution of Large-Scale Data Production Events in OpenStreetMap.” Transactions in GIS 25 2: 622–641. doi:10.1111/tgis.12746.
  • Gröchenig, S., R. Brunauer, and K. Rehrl. 2014. “Digging into the History of VGI Data-Sets: Results from a Worldwide Study on OpenStreetMap Mapping Activity.” Journal of Location Based Services 8 (3): 198–210. doi:10.1080/17489725.2014.978403.
  • Hecht, R., C. H. S. Kunze, and S. Hahmann. 2013. “Measuring Completeness of Building Footprints in OpenStreetMap Over Space and Time.” International Journal of Geo-Information 2 (4): 1066–1091. doi:10.3390/ijgi2041066.
  • Hua, Y., Z. Zhao, R. Li, X. Chen, Z. Liu, and H. Zhang. 2018. “Deep Learning with Long Short-Term Memory for Time Series Prediction.” IEEE Communications Magazine 57 (6): 114–119.
  • Juhász, L., and H. H. Hochmair. 2018. “Osm Data Import as an Outreach Tool to Trigger Community Growth? A Case Study in Miami.” ISPRS International Journal of Geo-Information 7 (3). doi:10.3390/ijgi7030113.
  • Juhász, L., T. H. H. Q. S. Novack, H. Hochmair, and S. Qiao. 2020. “Cartographic Vandalism in the Era of Location-Based Games—The Case of OpenStreetMap and Pokémon Go.” International Journal of Geo-Information 9: 197. doi:10.3390/ijgi9040197.
  • Lindemann, B., T. Müller, H. Vietz, N. Jazdi, and M. Weyrich. 2021. ”A Survey on Long Short-Term Memory Networks for Time Series Prediction.” Procedia CIRP 99: 650–655. 14th CIRP Conference on Intelligent Computation in Manufacturing Engineering, 15–17 July 2020. doi:10.1016/j.procir.2021.03.088.
  • Maguire, S., and M. Tomko. 2017. “Ripe for the Picking? Dataset Maturity Assessment Based on Temporal Dynamics of Feature Definitions.” International Journal of Geographical Information Science 31 (7): 1334–1358. doi:10.1080/13658816.2017.1287370.
  • Makridakis, S. 1993. “Accuracy Measures: Theoretical and Practical Concerns.” International Journal of Forecasting 9 (4): 527–529. doi:10.1016/0169-2070(93)90079-3.
  • Makridakis, S., and M. Hibon. 1979. “Accuracy of Forecasting: An Empirical Investigation.” Journal of the Royal Statistical Society 142 (2): 97–145.
  • Mocnik, F.-B., A. G. L. E. M. J. C. Mobasheri, and C. Klonner. 2018. “A Grounding-Based Ontology of Data Quality Measures.” Journal of Spatial Information Science 16: 1–25. doi:http://dx.doi.org/10.5311/JOSIS.2018.16.360.
  • Mocnik, F.-B., A. Zipf, and M. Raifer. 2017. “The OpenStreetMap Folksonomy and Its Evolution.” Geo-Spatial Information Science 20 (3): 219–230. doi:10.1080/10095020.2017.1368193.
  • Neis, P., and A. Zipf. 2008. “Zur kopplung von opensource, openls und openstreetmaps in openrouteservice.org.” AGIT - Symposium fuer Angewandte Geoinformatik, Salzburg, Austria.
  • Novack, T., Z. Wang, and A. Zipf. 2018. “A System for Generating Customized Pleasant Pedestrian Routes Based on OpenStreetMap Data.” Sensors 18: 3794. doi:10.3390/s18113794.
  • OpenStreetMap Contributors. 2017. “Planet dump.” https://planet.osm.org.https://www.openstreetmap.org
  • Raifer, M., R. Troilo, F. Kowatsch, M. Auer, L. Loos, S. Marx, and K. Przybill. 2019. “Oshdb: A Framework for Spatio-Temporal Analysis of OpenStreetMap History Data.” Open Geospatial Data, Software Stand 4 (1). doi:10.1186/s40965-019-0061-3.
  • Schultz, M., J. Voss, M. Auer, S. Carter, and A. Zipf. 2017. “Open Land Cover from OpenStreetMap and Remote Sensing.” International Journal of Applied Earth Observation and Geoinformation 63: 206–213. doi:10.1016/j.jag.2017.07.014.
  • Sehra, S., J. Singh, and H. Rai. 2017. “Using Latent Semantic Analysis to Identify Research Trends in OpenStreetMap.” ISPRS International Society for Photogrammetry and Remote Sensing 195: 6.
  • Senaratne, H., A. Mobasheri, A. L. Ali, C. Capineri, and M. M. Haklay. 2017. “A Review of Volunteered Geographic Information Quality Assessment Methods.” International Journal of Geographical Information Science 31 (1): 139–167. doi:10.1080/13658816.2016.1189556.
  • Sieber, R. E., and M. Haklay. 2015. “The Epistemology(s) of Volunteered Geographic Information: A Critique.” Geographical and Environmental 2 (2): 122–136.
  • Wang, X., K. Smith, and R. Hyndman. 2006. “Characteristic-Based Clustering for Time Series Data.” Data Mining and Knowledge Discovery 13 (3): 335–364. doi:10.1007/s10618-005-0039-x.
  • Xu, J., L. Li, and Q. Zhou. 2017. “Spatial-Temporal Analysis of OpenStreetMap Data After Natural Disasters: A Case Study of Haiti Under Hurricane Matthew.” ISPRS International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XLII-2/W7: 1401–1406.
  • Zhang, L., and D. Pfoser. 2019. “Using OpenStreetMap Point-Of-Interest Data to Model Urban Change - a Feasibility Study.” PLOS ONE. 14 (2): 1–34.
  • Zia, M. and Z. S. D. Cakir. “Turkey OpenStreetMap.”
  • Zook, M., M. Graham, T. Shelton, and S. Gorman. 2010. “Volunteered Geographic Information and Crowdsourcing Disaster Relief: A Case Study of the Haitian Earthquake.” World Medical & Health Policy 2 (2): 7–33. doi:10.2202/1948-4682.1069.

Absolute error statistics of the baseline forecasting method

The descriptive statistics of the absolute errors of the baseline forecasting method for horizons of 1 to 4 weeks are shown in below. The absolute error was calculated, as introduced, with the RMSE metric. For each of the four forecasting horizons, the statistics were computed for all 20 cities considered in this study and their respective test samples (see Section 3.5).

Table A1. Descriptive statistics of the absolute errors of the baseline method for a forecasting horizon of 1 week. The statistics were computed from all test samples from all 20 cities considered in this study.

Table A2. Descriptive statistics of the absolute errors of the baseline method for a forecasting horizon of 2 weeks. The statistics were computed from all test samples from all 20 cities considered in this study.

Table A3. Descriptive statistics of the absolute errors of the baseline method for a forecasting horizon of 3 weeks. The statistics were computed from all test samples from all 20 cities considered in this study.

Table A4. Descriptive statistics of the absolute errors of the baseline method for a forecasting horizon of 4 weeks. The statistics were computed from all test samples from all 20 cities considered in this study.