A Global Land Cover Training Dataset From 1984 To 2020: Data Descriptor
A Global Land Cover Training Dataset From 1984 To 2020: Data Descriptor
A Global Land Cover Training Dataset From 1984 To 2020: Data Descriptor
com/scientificdata
State-of-the-art cloud computing platforms such as Google Earth Engine (GEE) enable regional-
to-global land cover and land cover change mapping with machine learning algorithms. However,
collection of high-quality training data, which is necessary for accurate land cover mapping, remains
costly and labor-intensive. To address this need, we created a global database of nearly 2 million
training units spanning the period from 1984 to 2020 for seven primary and nine secondary land cover
classes. Our training data collection approach leveraged GEE and machine learning algorithms to ensure
data quality and biogeographic representation. We sampled the spectral-temporal feature space from
Landsat imagery to efficiently allocate training data across global ecoregions and incorporated publicly
available and collaborator-provided datasets to our database. To reflect the underlying regional class
distribution and post-disturbance landscapes, we strategically augmented the database. We used a
machine learning-based cross-validation procedure to remove potentially mis-labeled training units.
Our training database is relevant for a wide array of studies such as land cover change, agriculture,
forestry, hydrology, urban development, among many others.
1
Department of Earth and Environment, Boston University, 685 Commonwealth Avenue, Boston, MA, 02215,
USA. 2Department of Geosciences and Natural Resource Management (IGN), University of Copenhagen, DK-1350,
København K, Denmark. 3NASA Marshall Space Flight Center, Huntsville, AL, 35808, USA. 4Department of Natural
Resources and the Environment, University of Connecticut, Storrs, CT, 06269, USA. 5U.S. Geological Survey (USGS),
Earth Resources Observation and Science (EROS) Center, Sioux Falls, SD, 57198, USA. 6Imazon—Amazonia People
and Environment Institute, Belém, Brazil. 7Yale School of the Environment, Yale University, New Haven, CT, 06511,
USA. 8School of Biological Sciences, University of Utah, Salt Lake, UT, 84112, USA. 9Center for Remote Sensing and
Geographic Information Services, University of Ghana, Accra, Ghana. 10School of Life and Environmental Sciences,
Deakin University, Melbourne, Australia. 11Albo Climate, Ehad Ha’am, 9, Tel Aviv, Israel. 12Grupo de Investigación
de Biodiversidad, Medio Ambiente y Salud–BIOMAS, Universidad de las Américas (UDLA), Quito, Ecuador. 13Indigo
Ag, 500 Rutherford Avenue, Boston, MA, 02129, USA. 14REDD+ Coordination Unit, Oromia Environmental
Protection Authority, Addis Ababa, Ethiopia. 15Forestry Department Headquarters, Ministry of Green Economy and
Environment, Lusaka, Zambia. ✉e-mail: rkstan@bu.edu
global, and high-quality land cover training datasets that are adaptable to the needs of a wide range of end
users depending on the region of interest and the classification algorithm used. Currently, only a handful of
continental-to-global training14–18 and validation19,20 datasets are publicly available. Several large-scale bench-
mark remote sensing datasets, designed to support the development of deep learning algorithms targeting specific
applications, are publicly available (e.g., SpaceNet21, BigEarthNet22, and DeepSat23). While these datasets are val-
uable resources, the data collection efforts that produced them were uncoordinated and not standardized, largely
because community-wide best practices for training data collection are not well established (although, see2).
As a result, most of these datasets are limited by their geographic coverage, spatial resolution, observation den-
sity, time span, or quality.
The goal of the Global Land Cover Estimation (GLanCE) project is to provide high-quality long-term records
of land cover and land cover change at 30 m spatial resolution for the 21st century using the Landsat archive24.
As part of the GLanCE project, we present a new land cover training database that is designed to address the
limitations outlined above. In creating this database, we aim to provide global coverage, ensure accuracy of land
cover labels at 30 m spatial resolution, cover nearly four decades, and produce a geographically dense dataset.
Our training data collection and curation approach leverages relatively recent technological advances such as
cloud computing (e.g., Google Earth Engine (GEE)) and machine learning algorithms (e.g., Random Forest,
k-means clustering etc.) to enforce data quality and ecological representation. Specifically, we implement an iter-
ative quality assessment procedure that relies on expert review and a machine learning-based cross-validation
procedure to remove poorly labeled training data.
Given the global scope of GLanCE, combined with the time and resource-intensive nature of training data
collection, it was necessary to supplement the GLanCE data collection with external datasets and map prod-
ucts. Specifically, we harmonized seven publicly available land cover training datasets to be consistent with
GLanCE data and combined them into a global database14,16–20,25. Similarly, we harmonized and integrated
several collaborator-provided datasets26–28, along with datasets collected by Boston University team members
for various other projects29–34. Lastly, following numerous recent studies (e.g.,3,5,35,36), we sampled existing land
cover map products (i.e.,18,37,38) to fill geographic and thematic gaps in the dataset.
The objective of this paper is to describe the GLanCE training dataset, which is available to the public for use
in regional-to-global land cover and land cover change studies. The dataset is global, medium spatial resolution
(30 m), designed to be geographically and spectrally representative of all global ecoregions39, and spans the time
period between 1984 and 2020. It provides a harmonized, standardized, and comprehensive database with up
to 23 land cover characteristics per training unit. To support land cover change mapping, the dataset includes
up to 36 years (in select regions of the world) and, notably, includes information about abrupt and gradual land
cover change processes. It is designed to be global in scope but can be sub-sampled and adapted depending on
the study region of interest, the classification algorithm used, and the classification legend desired (e.g., broad
classes, land use, leaf type and phenology, etc.).
Methods
Training data collection. GLanCE training data were collected by a team of trained image analysts at Boston
University using the land cover key and a suite of online tools (https://github.com/parevalo/measures_collector,
using Google Earth Engine API). Image analysts interpreted land cover on-screen using a combination of
high-resolution Google Earth imagery, Landsat imagery, time series of spectral reflectance, vegetation indices,
and Landsat-derived Tasseled Cap transformations. In addition, image analysts used Google Earth photos and
StreetView (where available) to aid their interpretations. Hereafter we refer to each entry in the database, which
represents individual Landsat pixels, as a training unit. Each training unit corresponds to an interpretation by
an image analyst of Continuous Change Detection and Classification (CCDC) (https://developers.google.com/
earth-engine/datasets/catalog/GOOGLE_GLOBAL_CCDC_V1) time segments (explained below) (Fig. 1). Each
unit was assessed for quality and potentially flagged for review by a second image analyst. If one image analyst
disagreed with another on the land cover label of a given training unit, a third team member reviewed and, if nec-
essary, re-interpreted or removed the unit. Units were removed if there were no high-resolution imagery available
and team members had no way of determining the land cover with high confidence.
Over the course of data collection, the team of image analysts consisted of 6 to 12 members who were trained
to interpret satellite imagery for land cover attributes. Analysts alternated between interpreting sets of randomly
assigned training units and reviewing peers’ interpretations. All image analysts received the same training to
ensure consistency in their interpretation, followed by a practice training set for each continent that was col-
lectively discussed so that analysts learned from their errors and followed consistent interpretation protocols.
Training included image interpretation, guidance on how to use software tools that were developed to support
this activity40 (https://github.com/parevalo/measures_collector), use of ancillary data (Table 1), and class defi-
nitions (Table 2). In addition, quarterly refresher trainings and weekly meetings were conducted to provide
analysts with feedback regarding errors and inconsistencies in interpretations discovered during the review
process. As a final step, all training data were reviewed for clerical errors and compiled into unique databases
for each continent.
Because a single Landsat pixel can include multiple land cover types over time, an important component of
our training data collection protocol was the interpretation of land cover change. The core algorithm used in this
project, the CCDC algorithm41, operates on time series of satellite data to identify abrupt and transitional changes,
and stable land cover. The concept of a time segment—the period for which the intra- and inter-annual patterns
of surface reflectance can be described by a single model fit—is central to the CCDC algorithm. To illustrate this
concept, Fig. 1 shows a time series of shortwave infrared reflectance values for two Landsat pixels that both cor-
respond to forest at the beginning of their time series, which then transition to grasslands and slowly evolve back
Fig. 1 Continuous Change Detection and Classification (CCDC) model fits for two pixels in Europe that were
converted from forest to grassland, eventually transitioning back to forest. The time series plots show all Landsat
observations (points) in the Landsat SWIR1 band, and the CCDC model fits (lines). The high-resolution images
illustrate the land cover change on the ground. Map data: Google, Maxar Technologies.
to forests at the end of the time series. To confirm that various types of land-cover change are represented in the
training database, changes such as those exemplified in Fig. 1 were explicitly included in the dataset and labeled
according to the type of change (as described above). In the examples shown in Fig. 1, CCDC identifies distinct
time segments between detected change events on the land surface (shown in different colors); each segment is
assigned a single land cover label. The time segments representing stable land cover correspond to the subsets
of Landsat time series with relatively constant spectral and temporal reflectance patterns over several consecu-
tive years. In Fig. 1a, the forest was represented by a stable segment from 2000 to 2009. In contrast, transitional
segments correspond to land cover that gradually transforms over the time, where consecutive annual patterns
of surface reflectance change in a constant manner. For example, in Fig. 1b, forest regrowth from 2011 to 2019
is evident in the Landsat time series and identified by CCDC in the segment starting in 2011. Abrupt changes,
such as the logging event in 2009 in Fig. 1a, corresponded to high magnitude breaks in the CCDC time segments.
In the training dataset, stable and transitional labels were assigned based on visual interpretation of Landsat time
series, CCDC models, and reference high resolution imagery. We based our labels on the condition of the land
cover and on the slope of the time series.
Data collected by the GLanCE team. Image analysts interpreted training units from three sources: (1) the
System for Terrestrial Ecosystem Parameterization (STEP) training database18, (2) a sample generated via unsu-
pervised clustering of Landsat spectral-temporal features, and (3) a sample of feedback training units generated
to improve the accuracy of land cover classes or regions that were persistently misclassified. The STEP database,
Fig. 2 Training data key used to assign land cover attributes to training units. Green boxes show Level 1 land
cover classes and yellow boxes show Level 2 classes.
which was created to provide a representative sample of land cover from all global ecoregions, was created to sup-
port the MODIS Collection 6 Land Cover Type Product (MCD12Q1)18. To adapt the STEP database for GLanCE,
we randomly selected 10 Landsat pixels within each MCD12Q1 500 m pixel and visually interpreted them using
the procedure outlined in above and labeled them using the GLanCE land cover key (Fig. 2, Table 2). Only train-
ing units representing homogenous land cover were collected as training data; training units containing mixed
land covers were removed.
The second set of training data that image analysts collected was based on unsupervised clustering of
spectral-temporal features estimated from Landsat image time series using the CCDC algorithm. This approach
was stratified using the World Wildlife Fund (WWF) ecoregions39 to ensure that each ecoregion was adequately
represented in the training database. While STEP-based training data were designed to capture homogeneous
land cover, a key goal of cluster-based sampling was to locate and collect training units with more heterogeneous
land cover composition. Including units that represent heterogeneous land cover in the training dataset has been
shown to improve classification results and is an efficient way to collect training samples at continental-to-global
scales because a relatively small number of training units captures the variability in land cover spectral prop-
erties from each ecoregion3,4,6–8. As such, the cluster-based training data represented variation in land cover
composition, stability, and intra- and inter-class spectral variability at the regional scale.
Our cluster-based approach included two main steps: (1) principal component analysis (PCA) to reduce the
dimensionality of the data; and (2) k-means clustering on the principal components (PCs) to identify the opti-
mal partitioning of the training data. For each ecoregion, we selected a maximum of 10 PCs to capture at least
80% of the variance, although 99% of the variance was frequently captured in fewer PCs. We ran k-means clus-
tering for a range of K values from 5 to 400 and for each value we calculated the sum of squared distances from
each point to its assigned center to select the optimal (fewest) number of clusters that were well separated from
one another. As a result, we selected 30 to 60 clusters – for a maximum of 500 training units - per ecoregion. The
resulting dataset was representative of the distribution of land cover at the regional scale and included a mix of
homogeneous and heterogeneous training units. Because the STEP- and cluster-based datasets were collected
explicitly for the purposes of the GLanCE project, together they represent the most thematically detailed and
complete data in our database, with up to 23 recorded attributes per training unit.
The third set of training data were generated to iteratively improve the accuracy of land cover maps. Despite
our best efforts to represent all ecoregions in the training data, examples of some specialized and regionally rel-
evant land uses (e.g., greenhouses in Spain and China, sparse orchards, and plantations in India etc.) were miss-
ing in our training database, which resulted in obvious errors in the map results. To ameliorate these issues, we
collected “Feedback” training units for these locations around the globe using the interpretation tools described
above.
Supplementary data sources. Given the global scale of the GLanCE project and the lack of avail-
able high-resolution imagery in some parts of the world, the GLanCE-collected dataset described above was
insufficient to capture the full range of geographic, temporal, and spectral heterogeneity in global land cover.
To address this, we supplemented the GLanCE training database by leveraging existing publicly available,
collaborator-contributed, and team-collected datasets that we harmonized and standardized to conform to the
GLanCE land cover classification key. The sources and key features of these data are summarized in Table 1.
As part of this process, we worked with collaborators and team members with regional expertise in land cover
and land cover change to harmonize their land cover legend with the GLanCE legend. This consisted of cross
walking collaborator-provided land cover legends and definitions to the GLanCE legend, clarifying definitions
with collaborators, and eliminating unsuitable classes from the pool of potentially useful training units. Data
labeled as “BU team collected” were compiled from different existing projects by team members but weren’t
ASB crop Aral Sea Basin 2008, 2011, 2015, 2016, 2017, 2018 6,202 25
Colombia 2001–2017* 31
Australia 1985–2019 26
Table 1. Summary of data sources used to compile the GLanCE training database. STEP stands for System for
Terrestrial Ecosystem Parameterization; ABoVE stands for Arctic Boreal Vulnerability Experiment; LCMAP
stands for Land Change Monitoring, Assessment, and Projection; LUCAS stands for Land Use and Coverage
Area frame Survey; GLC30 stands for Global Land Cover product with Fine Classification System; ASB stands
for Aral Sea Basin. *Collected to align with CCDC model fits.
explicitly collected for the GLanCE project (Table 1). Note that several of these data sets are publicly availa-
ble and peer-reviewed, and all datasets were extensively vetted for quality control as described below. While
GLanCE-collected training units were collected based on CCDC segments, most supplementary data sources
were not.
Unfortunately, even after the data collected by the GLanCE team were combined with the supplementary
datasets described above, some land cover classes, especially rarer classes (e.g., developed, water, shrub), were
underrepresented. To address this, we augmented the database with training data derived from the World
Settlement Footprint product37, the Global Surface Water product38, and by implementing the algorithm devel-
oped by Zhang & Roy (2017)36 that uses the MODIS Land Cover Type Product18 to automatically select can-
didate training units at Landsat resolution. Briefly, this algorithm estimates the 20th, 50th, and 80th percentile
for thirteen different Landsat variables (bands and band ratios) for a total of 39 predictors, calculates the metric
centroid (a vector of 39 metric average values) of all 30 m pixels located within 500 m MODIS pixels, and then
selects the 30 m Landsat pixel with the smallest absolute difference from the metric centroid. Since the World
Settlement Footprint and Global Surface Water products are produced at spatial resolutions that are comparable
to Landsat, we drew a random sample of training units for each ecoregion from each product in each conti-
nent. We used these units to ensure that the training data were representative of the underlying distribution
of land cover, both regionally and globally, which is a common problem in machine learning-based land cover
classification6,42.
Table 2. GLanCE Level 1 and 2 land cover legend definitions. The integer values assigned to land cover classes
are indicated in parentheses for Level 1 and Level 2 labels.
For example, Step 4 occasionally resulted in supplementary datasets being reviewed and reinterpreted by GLanCE
image analysts. Details regarding legend harmonization for each data set are included in the Supplementary
Information. Despite our best efforts to harmonize and clean the supplementary training datasets, a constraint to
our approach is that we had limited control over the external dataset’s accuracy and consistency.
Data Records
The GLanCE training dataset described in this paper is available from Source Cooperative45 under Creative
Commons license CC-BY-4.0.
The GLanCE land cover training dataset includes two nested sets of classes: seven broad, mutually exclusive
classes (Level 1) and nine secondary classes (Level 2) (Table 2, Fig. 2)24. The GLanCE land cover classification
scheme is designed to focus primarily on land cover and is compatible with common land use categories for
greenhouse gas inventory reporting46 land cover classification systems such as the IPCC and the FAO Land
Cover Classification System (LCCS)24,47. Land cover labels were assigned for 30 m Landsat training units based
on fractional cover thresholds (Fig. 2). To label each training unit, we followed the stepwise decision-making
process shown in Fig. 2. In addition to Level 1 & 2 land cover and land use labels, the dataset includes eight
additional attributes that provide complementary information related to land cover and land use properties
(Table S1). Note that these additional attributes (e.g., LC_Confidence, Segment_Type) were collected only
for GLanCE training units and as a result are available only for ~3% of the dataset. Each training unit’s land
cover label corresponds to a specific Landsat training unit and time period between 1984 and 2020 (Table S1).
The years assigned to each unit are inclusive. For example, if the training unit is labeled “Trees” from 2000 to 2020,
then it was trees for the full calendar years of 2000 and 2020. IDs are assigned based on latitude and longitude,
so units with duplicate location and the samet ID indicate units that experienced land cover abrupt or gradual
change. However, each training unit has a unique ‘Glance_ID’. Table S1 includes a complete list of attributes and
descriptions.
Because land cover is dynamic and can change due to natural or anthropogenic processes, GLanCE training
units are characterized as either ‘stable’ or ‘transitional’ (Segment_Type in Table S1) based on time series of
both high-resolution imagery and Landsat observations (detailed description in Methods). A “stable” unit is
defined as a time segment with a single land cover and consistent annual patterns of spectral reflectance across
the length of the segment. Conversely, a “transitional” unit is a time segment in which the land cover gradually
changes over the time, where the transition is reflected in the annual patterns of surface reflectance. Transitional
units included examples of forest decline or degradation (Fig. 1), fire, insect damage, urban development, flood-
ing, regrowth (Fig. 1), riparian/water shift, drought, as well as unknown transitions. For transitional processes
that slowly transform landscapes, such as vegetation regrowth during ecological succession, the land cover was
recorded at both the beginning and end of the time segment (typically the first and last three years) when the
Level 1 land cover attributes remain unchanged (Fig. 1).
The V1.0 training dataset consists of 1,874,995 training units distributed across the globe representing seven
broad land cover classes at Level 1 and nine classes at Level 2. Approximately 79% and 21% of the dataset
Water 183,160
Herbaceous
655,076
Trees
921,564
Developed 48,185
Fig. 3 GLanCE global training units class distribution for Level 1 land cover. Note that the Bare label is
shorthand for Barren/sparsely vegetated.
correspond to stable land cover and change, respectively. Note that the change category includes abrupt (land
cover label change) and transitional processes such forest regrowth, coastal water dynamics etc. At global scale,
change is a relatively rare occurrence but because it is inherently hard to map, change needs to be captured in
the training database. Figures 3, 5a show the class frequency and geographic distribution of the training data-
base for Level 1 land cover, while Figs. 4, 5b display the distribution for Level 2 land cover and land use. Level
1 training data are well distributed and representative of all ecoregions. However, only 50% of the training data
contain Level 2 legend information (Figs. 4, 5). Despite our efforts to include Level 2 labels for supplementary
datasets whenever possible, only half of the training units contain this information in the final training database
because supplementary datasets often did not include land use (e.g., agriculture). Relatively rare classes on the
landscape such as developed, barren/sparsely vegetated and shrubs are well-represented in the training database.
In contrast, ice/snow is not well represented because it tends to be located in areas where Landsat data density
is insufficient for CCDC (Fig. 3). The distribution of Level 2 labels is dominated by herbaceous classes such as
grasslands and agriculture (Figs. 4, 5b). The detailed quality control procedures, described in the Methods sec-
tion, were applied to Level 1 land cover labels only. Therefore, Level 2 labels and all other land cover attributes
(e.g., leaf phenology) may not always meet the highest standards of quality and may need to be further filtered
or processed to fit the user’s region and research of interest.
To our knowledge, the dataset presented in this study is the longest, most extensive, and comprehensive pub-
licly available global land cover and land use training database. We standardized and harmonized 22 disparate
sources of land cover labels into a single unified training database that is comprised of 39% publicly available
data, 55% collaborator-provided data, 4% GLanCE-collected data (collected explicitly for the purposes of the
GLanCE product), 1% Boston University team collected data (collected by team members for other projects, not
explicitly for the purposes of GLanCE), and 0.2% MODIS-derived training data (Table 1, Fig. 6b). Among the
various data sets incorporated into the database, the GLanCE training data contain the highest level of detailed
ancillary information on secondary land cover attributes and change information, and frequently span 20 years
between 1999 and 2019 (Fig. 6b). Some collaborator-provided data sources such as MapBiomas, LCMAP, and
ABoVE (Table 1) include up to 35 years of land cover labels and change information, while most publicly avail-
able data were limited to a single year (Fig. 6b). The training units are not equally distributed among continents
(Fig. 6a) with Europe (source: LUCAS) and South America (source: MapBiomas) having the most data available.
While training units in South America have time segment lengths of up to 35 years, most training units have
time segment lengths of less than 7 years (Fig. 6c). Level 1 land cover classes are evenly distributed among all the
time segment length bins (Fig. 6d).
Technical Validation
Human error is inherent to all land cover training data sets, especially those compiled by on-screen interpre-
tation2. However, the size of our database precludes manual quality assurance for each training unit. Hence, to
minimize errors and maximize the dataset’s quality, we used a two-step machine learning-based cross-validation
procedure adapted from Brodley & Friedl48 to remove poorly labeled training data. In the first step, we divided
the training data in each continent into up to 9 biogeographical regions determined based on k-means cluster-
ing of maximum temperature, minimum temperature, precipitation, latitude, and longitude. Before clustering,
Soil 2,056
Deciduous 128,094 Evergreen 135,114 Mixed 91,342
Rock 9,972
Fig. 4 GLanCE global training data distribution for Level 2 land cover. Note that this figure doesn’t display the
Moss/lichen category because it includes only 640 training units.
Fig. 5 Global spatial distribution of GLanCE training units for Level 1 (a) and Level 2 (b) land cover. Note
that the overall density of training units is lower in Africa and Asia, so the Developed class in (a) is visually
overrepresented in this map.
Fig. 6 The distribution of training units across the different continents (a) and across time (b–d). Panels b
thru d show the number of training units that belong to different bins of time segment lengths split by data
source (b), continent (c), and land cover class (d). BU stands for Boston University. MODIS-derived units
accounted for 0.2% of the total dataset so they are not shown in panel (b). SAmerica stands for South America
and NAmerica stands for North America. The insets show more detail for the bins of time segment lengths with
fewer observations.
the three climate variables were normalized to unit variance, and we included latitude and longitude to create
spatially coherent clusters. In the second step, we used the ranger package in R to estimate unique random forest
models for each cluster and to estimate class membership probabilities for each class for each training unit using
a combination of remote sensing and ancillary features as predictors at each unit (see24 for details). The ancillary
data features used to build random forest models include topography, global surface water, climate, population,
and settlements.
We then examined the difference between the 1st and 2nd most likely classes; training data with margins less
than 0.05 were discarded because they represented cases where the two most likely classes were easily confused.
To select this threshold, we performed a 10-fold cross validation analysis, which demonstrated that training
cases with margins less than or equal to 0.05 had substantially higher misclassification rates relative to cases with
higher margins. We then removed all misclassified cases for which the margin between the predicted label and
the label assigned to the unit in the database was in the upper quartile of margins for each class. In other words,
we removed data where the assigned label differed from the label predicted by random forest, and where the
class probability for the label assigned by random forest was high.
Using this procedure, we removed ~15% of the training data in each continent (Fig. 7 - removed data shown
in gray). Approximately 97% of removed training units were discarded by applying the 0.05 threshold and 3%
were discarded because the predicted and assigned label differed. The removal of training units was uniform
across land covers (Fig. 7) and did not disproportionately affect transitional units (our procedure discarded 23%
of all training units labeled transitional and 16% of all training units labeled stable). In some cases, training units
with very small margins provide examples of land cover that are unique or represent important sub-variants
of the classes we are mapping. In general, however, these cases represent training units with mixed land cover
and the small margins associated with these training units indicate that the classifier is not able to reliably
Fig. 7 Class distribution before and after the filtering procedure. Gray bars show the removed units per class
based on the cross-validation procedure. Note that the Bare label is shorthand for Barren/sparsely vegetated.
The Ice/snow class contained only 214 units and is not included in this visual.
distinguish their land cover class, which can increase errors in classification results48. Hence, we adopted a strat-
egy of removing these cases from the training set.
As an additional technical validation, we followed an approach used by Doda et al.49 to compare classification
results based on our training data against reference data. We created the reference set by splitting our final training
database into train (70%) and test (30%) data and withholding the test data from the model. We used the random
forest classification algorithm in the scikit-learn package (version 0.22.2) in Python 3.6.7. We selected random
forest because it is 1) a widely-used model suitable for land cover mapping, 2) relatively resistant to overfitting,
and 3) efficiently handles noisy input data50,51. To optimize random forest, we used a grid search to automatically
fine-tune model parameters and 3-fold cross-validation to assess model performance. To estimate the final model
used in this technical validation, we used the parameter combination with the lowest root-mean-squared error
(RMSE). To train the model, we used predictors derived from the CCDC parameters based on harmonic models
fit to time series of Landsat surface reflectance bands (e.g., green phase, green amplitude, etc.) as well as a variety
of ancillary layers (e.g., topography, population, etc.)24. In total, each training unit had 56 features.
Tables S2–S7 show the confusion matrices between the observed and predicted land cover labels. Even
though the user’s accuracy is high (greater than 0.8 in most continents) (Fig. 8a), there is confusion between
some classes (Tables S2–S7). Distinguishing between herbaceous vegetation and shrubs is difficult because fre-
quently training units represent a mixture of these two cover types, an issue that has been well documented by
previous studies18,24. Even at 30 m spatial resolution, relatively few training units are actually uniform in terms
of their land cover composition. Note that producer’s accuracy and F1-score are generally high for most classes,
but low for shrubs for almost every continent except for Oceania (Fig. 8b,c). Producer’s accuracy is also low for
developed and bare ground, especially in South America (Fig. 8b).
Usage Notes
Because the process of acquiring supplementary datasets was opportunistic and non-systematic based on data
availability and quality, the full database includes geographic variation in data density. For example, some
regions have training units that are geographically clumped (e.g., Ghana) or land cover classes that are overrep-
resented (e.g., herbaceous) (Figs. 3–5). Some users may need to sub-sample the dataset or enforce constraints on
data density depending on their research question, application, or area of interest. For example, in the GLanCE
project, we use a 100 × 100 km grid to assign weights such that if more training data are present in a single grid
cell, the data are sub-sampled (and vice-versa) to ensure pseudo-uniform density of training data across space.
Users may also need to sub- or resample the database to enforce uniform or proportional class distributions.
For applications focused on land cover change (abrupt or gradual), for which our database includes propor-
tionally less data, we recommend retaining all change training data (for guidance see3,11,13). For applications
focused on agriculture, users can use the Level 2 category of ‘Agriculture’ as a starting point but should be aware
that this label has not undergone rigorous quality control and is available for a limited subset of the global
training dataset (286,284 units). Users should apply quality control measures, depending on their applications,
such as filtering Level 2 attributes as outlined in the Technical Validation, intersecting the agriculture units with
existing regional cropland maps as outlined in Methods, and visual inspection of training units in Google Earth.
Code availability
We used open-source tools to ensure transparency and reproducibility of our research, including R (4.3.0), Python
3.6.7, and Google Earth Engine. Time series tools for training data collection are available on GitHub (https://
github.com/parevalo/measures_collector) as is the repository for filtering training data (https://github.com/ma-
friedl/GlanceFiltering). Custom continental definitions can be found at this repository: https://measures-glance.
Fig. 8 User’s accuracy (a), producer’s accuracy (b), and F1 score (c) for each continent and each land cover class
(except Ice/Snow as there weren’t enough training units available for this class). N America stands for North
America, and S America for South America.
github.io/glance-grids/params. Continuous Change Detection and Classification (CCDC) tools and applications
can be found on Google Earth Engine (https://glance.earthengine.app/view/fulltstools) and python (https://
github.com/repository-preservation/lcmap-pyccd).
References
1. Foody, G., Pal, M., Rocchini, D., Garzon-Lopez, C. & Bastin, L. The sensitivity of mapping methods to reference data quality:
Training supervised image classifications with imperfect reference data. ISPRS Int. J. Geo-Inf. 5, 199 (2016).
2. Elmes, A. et al. Accounting for training data error in machine learning applied to earth observations. Remote Sens. 12, 1034 (2020).
3. Colditz, R. An evaluation of different training sample allocation schemes for discrete and continuous land cover classification using
decision tree-based algorithms. Remote Sens. 7, 9655–9681 (2015).
4. Foody, G. M. & Mathur, A. The use of small training sets containing mixed pixels for accurate hard image classification: Training on
mixed spectral responses for classification by a SVM. Remote Sens. Environ. 103, 179–189 (2006).
5. Zhou, Q., Tollerud, H., Barber, C., Smith, K. & Zelenak, D. Training data selection for annual land cover classification for the Land
Change Monitoring, Assessment, and Projection (LCMAP) Initiative. Remote Sens. 12, 699 (2020).
6. Maxwell, A. E., Warner, T. A. & Fang, F. Implementation of machine-learning classification in remote sensing: an applied review. Int.
J. Remote Sens. 39, 2784–2817 (2018).
7. Mountrakis, G., Im, J. & Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 66,
247–259 (2011).
8. Rodriguez-Galiano, V. F., Ghimire, B., Rogan, J., Chica-Olmo, M. & Rigol-Sanchez, J. P. An assessment of the effectiveness of a
random forest classifier for land-cover classification. ISPRS J. Photogramm. Remote Sens. 67, 93–104 (2012).
9. Frenay, B. & Verleysen, M. Classification in the presence of label noise: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 25, 845–869 (2014).
10. Huang, C., Davis, L. S. & Townshend, J. R. G. An assessment of support vector machines for land cover classification. Int. J. Remote
Sens. 23, 725–749 (2002).
11. Jin, H., Stehman, S. V. & Mountrakis, G. Assessing the impact of training sample selection on accuracy of an urban classification: a
case study in Denver, Colorado. Int. J. Remote Sens. 35, 2067–2081 (2014).
12. Lu, D. & Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote
Sens. 28, 823–870 (2007).
13. Zhu, Z. et al. Optimizing selection of training and auxiliary data for operational land cover classification for the LCMAP initiative.
ISPRS J. Photogramm. Remote Sens. 122, 206–221 (2016).
14. Alemohammad, H. & Booth, K. LandCoverNet: A global benchmark land cover classification training dataset. ArXiv201203111 Cs (2020).
15. Laso Bayas, J. C. et al. A global reference database of crowdsourced cropland data collected using the Geo-Wiki platform. Sci. Data
4, 170136 (2017).
16. d’Andrimont, R. et al. Harmonised LUCAS in-situ land cover and use database for field surveys from 2006 to 2018 in the European
Union. Sci. Data 7, 352 (2020).
17. Fritz, S. et al. A global dataset of crowdsourced land cover and land use reference data. Sci. Data 4, 170075 (2017).
18. Sulla-Menashe, D., Gray, J. M., Abercrombie, S. P. & Friedl, M. A. Hierarchical mapping of annual global land cover 2001 to present:
The MODIS Collection 6 Land Cover product. Remote Sens. Environ. 222, 183–194 (2019).
19. Liangyun, L., Yuan, G., Xiao, Z., Xidong, C. & Shuai, X. A dataset of global land cover validation samples, Zenodo, https://doi.
org/10.5281/zenodo.3551995 (2019).
20. Pengra, B. W. et al. LCMAP reference data product 1984–2018 land cover, land use and change process attributes. https://doi.
org/10.5066/P9ZWOXJ7 (2020).
21. Van Etten, A., Lindenbaum, D. & Bacastow, T. M. SpaceNet: A remote sensing dataset and challenge series. Preprint at http://arxiv.
org/abs/1807.01232 (2019).
22. Sumbul, G., Charfuelan, M., Demir, B. & Markl, V. BigEarthNet: A large-scale benchmark archive for remote sensing image
understanding. in IGARSS 2019 - 2019 IEEE International Geoscience and Remote Sensing Symposium 5901–5904, https://doi.
org/10.1109/IGARSS.2019.8900532 (2019).
23. Basu, S. et al. DeepSat - A Learning framework for satellite imagery. Preprint at http://arxiv.org/abs/1509.03602 (2015).
24. Friedl, M. A. et al. Medium spatial resolution mapping of global land cover and land cover change across multiple decades from
Landsat. Front. Remote Sens. 3, 894571 (2022).
25. Remelgado, R. et al. A crop type dataset for consistent land cover classification in Central Asia. Sci. Data 7, 250 (2020).
26. Calderón-Loor, M., Hadjikakou, M. & Bryan, B. A. High-resolution wall-to-wall land-cover mapping and land change assessment
for Australia from 1985 to 2015. Remote Sens. Environ. 252, 112148 (2021).
27. Souza, C. M. et al. Reconstructing three decades of land use and land cover changes in Brazilian biomes with Landsat archive and
Earth Engine. Remote Sens. 12, 2735 (2020).
28. CILSS. Landscapes of West Africa - A WindoW on A ChAnging World. (2016).
29. Chen, S. et al. Monitoring temperate forest degradation on Google Earth Engine using Landsat time series analysis. Remote Sens.
Environ. 265, 112648 (2021).
30. Chen, S., Olofsson, P., Saphangthong, T. & Woodcock, C. E. Monitoring shifting cultivation in Laos with Landsat time series. Remote
Sens. Environ. 288, 113507 (2023).
31. Arévalo, P., Olofsson, P. & Woodcock, C. E. Continuous monitoring of land change activities and post-disturbance dynamics from
Landsat time series: A test methodology for REDD+ reporting. Remote Sens. Environ. 238, 111051 (2020).
32. Graesser, J. et al. Temporally-consistent annual land cover from Landsat time series in the Southern Cone of South America. Remote
Sens. 14, 4005 (2022).
33. Stanimirova, R., Graesser, J., Olofsson, P. & Friedl, M. A. Widespread changes in 21st century vegetation cover in Argentina,
Paraguay, and Uruguay. Remote Sens. Environ. 282, 113277 (2022).
34. Wang, J. A. et al. Extensive land cover change across Arctic–Boreal Northwestern North America from disturbance and climate
forcing. Glob. Change Biol. 26, 807–822 (2020).
35. Radoux, J. et al. Automated training sample extraction for global land cover mapping. Remote Sens. 6, 3965–3987 (2014).
36. Zhang, H. K. & Roy, D. P. Using the 500 m MODIS land cover product to derive a consistent continental scale 30 m Landsat land
cover classification. Remote Sens. Environ. 197, 15–34 (2017).
37. Marconcini, M. et al. Outlining where humans live, the World Settlement Footprint 2015. Sci. Data 7, 242 (2020).
38. Pekel, J.-F., Cottam, A., Gorelick, N. & Belward, A. S. High-resolution mapping of global surface water and its long-term changes.
Nature 540, 418–422 (2016).
39. Olson, D. M. et al. Terrestrial ecoregions of the world: A new map of life on earth. BioScience 51, 933 (2001).
40. Arévalo, P., Bullock, E. L., Woodcock, C. E. & Olofsson, P. A suite of tools for continuous land change monitoring in Google Earth
Engine. Front. Clim. 2, 576740 (2020).
41. Zhu, Z. & Woodcock, C. E. Continuous change detection and classification of land cover using all available Landsat data. Remote
Sens. Environ. 144, 152–171 (2014).
42. Mellor, A., Boukir, S., Haywood, A. & Jones, S. Exploring issues of training data imbalance and mislabelling on random forest
performance for large area land cover classification using the ensemble margin. ISPRS J. Photogramm. Remote Sens. 105, 155–168 (2015).
43. Zanaga, D. et al. ESA WorldCover 10 m 2020 v100, Zenodo, https://doi.org/10.5281/zenodo.5571936 (2021).
44. Buchhorn, M. et al. Copernicus global land cover layers—Collection 2. Remote Sens. 12, 1044 (2020).
45. Stanimirova, R. et al. A global land cover training dataset from 1984 to 2020, Radiant MLHub, Version 1.0, https://doi.org/10.34911/
rdnt.x4xfh3 (2023).
46. GFOI. Integration of remote-sensing and ground-based observations for estimation of emissions and removals of greenhouse gases
in forests: Methods and Guidance from the Global Forest Observations Initiative. (2020).
47. Di Gregorio, A. & Jansen, L. J. M. Lands cover classification system (LCCS). (2000).
48. Brodley, C. E. & Friedl, M. A. Identifying mislabeled training data. J. Artif. Intell. Res. 11, 131–167 (1999).
49. Doda, S. et al. So2Sat POP - A curated benchmark data set for population estimation from space on a continental scale. Sci. Data 9,
715 (2022).
50. Belgiu, M. & Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm.
Remote Sens. 114, 24–31 (2016).
51. Breiman, L. Random Forests. Mach. Learn. 45, 5–32 (2001).
Acknowledgements
This research was supported by the NASA Making Earth System Data Records for Use in Research Environments
(MEaSUREs) program, grant number 80NSSC18K0994. Any use of trade, firm, or product names is for
descriptive purposes only and does not imply endorsement by the U.S. Government.
Author contributions
Conceptualization: K.T., K.T., R.S., M.A.F., C.E.W., K.M., S.S.; Methodology & Formal analysis: K.T., K.T., R.S.,
M.A.F., K.M., S.S., K-T.H.; Software: P.A., E.B., K-T.H., M.A.F., R.S., K.T., Y.Z.; Validation: M.A.F., R.S.; Resources:
M.A.F., C.E.W., C.P.B., P.A., E.B., Z.Z., P.O., C.S., S.C., J.A.W., F.M., M.C-L., M.H., B.A.B., J.G., D.L.B., B.M., S.S.,
A.S.; Data Curation: K.M., K.T., K.T., R.S., S.S., K-T.H.; Visualization: R.S.; Supervision: M.A.F., C.E.W., Z.Z., P.O.,
C.P.B.; Writing - Original Draft: R.S., K.T., K.T.; Writing - Review & Editing: R.S., K.T., K.T., P.O., Z.Z., M.A.F., C.S.
Competing interests
The authors declare no competing interests.
Additional information
Supplementary information The online version contains supplementary material available at https://doi.org/
10.1038/s41597-023-02798-5.
Correspondence and requests for materials should be addressed to R.S.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Cre-
ative Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not per-
mitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the
copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.