the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
TreeSatAI Benchmark Archive: a multi-sensor, multi-label dataset for tree species classification in remote sensing
Steve Ahlswede
Christian Schulz
Christiano Gava
Patrick Helber
Benjamin Bischke
Michael Förster
Florencia Arias
Jörn Hees
Begüm Demir
Birgit Kleinschmit
Airborne and spaceborne platforms are the primary data sources for large-scale forest mapping, but visual interpretation for individual species determination is labor-intensive. Hence, various studies focusing on forests have investigated the benefits of multiple sensors for automated tree species classification. However, transferable deep learning approaches for large-scale applications are still lacking. This gap motivated us to create a novel dataset for tree species classification in central Europe based on multi-sensor data from aerial, Sentinel-1 and Sentinel-2 imagery. In this paper, we introduce the TreeSatAI Benchmark Archive, which contains labels of 20 European tree species (i.e., 15 tree genera) derived from forest administration data of the federal state of Lower Saxony, Germany. We propose models and guidelines for the application of the latest machine learning techniques for the task of tree species classification with multi-label data. Finally, we provide various benchmark experiments showcasing the information which can be derived from the different sensors including artificial neural networks and tree-based machine learning methods. We found that residual neural networks (ResNet) perform sufficiently well with weighted precision scores up to 79 % only by using the RGB bands of aerial imagery. This result indicates that the spatial content present within the 0.2 m resolution data is very informative for tree species classification. With the incorporation of Sentinel-1 and Sentinel-2 imagery, performance improved marginally. However, the sole use of Sentinel-2 still allows for weighted precision scores of up to 74 % using either multi-layer perceptron (MLP) or Light Gradient Boosting Machine (LightGBM) models. Since the dataset is derived from real-world reference data, it contains high class imbalances. We found that this dataset attribute negatively affects the models' performances for many of the underrepresented classes (i.e., scarce tree species). However, the class-wise precision of the best-performing late fusion model still reached values ranging from 54 % (Acer) to 88 % (Pinus). Based on our results, we conclude that deep learning techniques using aerial imagery could considerably support forestry administration in the provision of large-scale tree species maps at a very high resolution to plan for challenges driven by global environmental change. The original dataset used in this paper is shared via Zenodo (https://doi.org/10.5281/zenodo.6598390, Schulz et al., 2022). For citation of the dataset, we refer to this article.
- Article
(5842 KB) - Full-text XML
- BibTeX
- EndNote
1.1 Importance of aerial imagery for forest monitoring
Public and private forest owners in Europe are increasingly confronted with adverse long- and short-term effects on forests driven by global environmental change (IPCC, 2014; MacDicken et al., 2016). The monitoring of damages caused by events such as storms, heat waves, disease outbreaks and insect infestations has become a highly prominent topic in Earth observation (Holzwarth et al., 2020), which is reflected in the recent remote sensing literature (Tanase et al., 2018; Hollaus and Vreugdenhil, 2019; Senf et al., 2020; Schuldt et al., 2020; Kowalski et al., 2020; Thonfeld et al., 2022). Thus, long-term mitigation strategies such as the conversion from mono-cultures to more diverse tree stands with higher resilience to environmental pressures have become highly relevant (Hlásny et al., 2017). To support these strategies, tree species that are most sensitive to recent and upcoming changes must be cost-efficiently mapped and monitored on a large scale.
To date, aerial imagery (i.e., digital orthophotographs) taken by plane is the primary large-scale remote sensing data source for forest authorities in Germany. Aerial imagery allows for fieldwork, which is time-consuming and expensive, to be reduced and planned more efficiently. With a spatial resolution in the decimeter range, it also constitutes a considerable source of data for visual check-ups and updates of forest management plans. However, individual species determination through visual interpretation is labor-intensive and requires specialized knowledge by the viewer (Hamedianfar et al., 2022).
1.2 State of research
In the field of forest remote sensing, aerial imagery is less frequently used than satellite imagery, especially when it comes to forest-type classification tasks (Holzwarth et al., 2020). This may be driven by the low number of spectral bands (red, green, blue, and near-infrared – RGB + NIR), which is considered to be a limitation in remote sensing applications (Fassnacht et al., 2016; Ganz et al., 2020). On the other hand, many studies have shown the strengths of multi-spectral satellite imagery for forest classification tasks (Pasquarella et al., 2018; Grabska et al., 2019; Immitzer et al., 2019; Ottosen et al., 2020; Hemmerling et al., 2021; Kollert et al., 2021; Waser et al., 2021; Welle et al., 2022). The extensive use of the spaceborne data with a much coarser resolution may be boosted by the open data policies of the Landsat and Sentinel programs. However, a higher spatial resolution is an essential factor for improving species prediction (Xu et al., 2021). For example, on a small spatial extent, studies using RGB imagery from unmanned aerial vehicles with resolutions in the centimeter range have resulted in very high prediction accuracies for species classification using deep learning (DL) techniques (Kattenborn et al., 2020, 2021; Schiefer et al., 2020). On a large spatial extent, studies using aerial imagery with a resolution in the decimeter range show its potential for forest cover mapping (Ganz et al., 2020) and foliage type classification (Krzystek et al., 2020) while integrating additional data from lidar and multi-spectral sensors. However, few studies have focused on species classification on aerial imagery (Holzwarth et al., 2020).
In summary, the state of research shows the general benefit of multi-sensor and multi-spectral datasets for forest species classification on all scales. Very high-resolution RGB imagery in the centimeter range has also successfully been used for species classification. However, the applicability of similar methods at a larger spatial extent remains unclear. This research gap motivated us to create a benchmark dataset for tree species mapping in central Europe based on the commonly available aerial imagery and additional freely accessible sensor sources.
1.3 Deep learning for large-scale forest applications
Traditional classifiers such as random forest (Breiman, 2001) or support vector machines (Boser et al., 1992) have been commonly used for classifying tree species using multi-spectral images (Sesnie et al., 2010; Karlson et al., 2015). One drawback of such approaches is their reliance on handcrafted features (i.e., expert rules) which require domain knowledge and often fail to include all pertinent features (Wurm et al., 2019). DL-based models are able to overcome this requirement as they automatically learn large sets of features which are specific to the given classification targets. Thus, DL models have attracted great attention for tree species classification (Egli and Höpke, 2020; Schiefer et al., 2020; Martins et al., 2021; Zhang et al., 2021). However, DL models require large amounts of labeled data in order to sufficiently learn the optimal model parameters, leading to bottlenecks in their application (Hamedianfar et al., 2022). One possible solution to this problem is to use networks which have been pre-trained on large datasets such as ImageNet (Deng et al., 2009) and then train the network on the smaller target training set afterwards. While this approach does indeed improve performance over training a model from scratch on small training datasets, recent work has shown that pre-training the network on imagery which is more closely related to the target dataset leads to a better performance (Sumbul et al., 2019).
DL datasets for large-scale land use and land cover applications often provide forests as a single class (e.g., UC Merced, Yang and Newsam, 2010; Deepsat, Basu et al., 2015) and EuroSAT (Helber et al., 2019) or as multiple classes based on foliage, seasonality and density types (e.g., BigEarthNet, Sumbul et al., 2019; SEN12MS, Schmitt et al., 2019). The NeonTreeCrowns (Weinstein et al., 2021) and NEON Tree Evaluation datasets (Weinstein et al., 2019a) represent the first large-scale DL datasets exclusively related to trees. They include data from multiple sensors including lidar, RGB, aerial, and hyperspectral imagery across different forest types in the United States. With DeepForest, the authors also proposed a semi-supervised DL neural network for individual tree crown delineation using airborne RGB data (Weinstein et al., 2019a, b). However, DL datasets for tree-species-level classification on a large scale have not yet been published. Therefore, we believe that the research community could greatly benefit from a dataset upon which to pre-train DL models, given that the model will have already learned features relevant for tree species.
1.4 Study aim
Given the potential improvement in performance from high-resolution aerial imagery and additional spectral information from Sentinel sensors, we introduce the TreeSatAI Benchmark Archive (Schulz et al., 2022). Based on reference data from forest administration data in Germany, the dataset aims to gather multi-sensor and multi-label information for the classification of 20 tree species in central Europe. The TreeSatAI Benchmark Archive consists of 50 381 image patches from aerial, Sentinel-2 (S2), and Sentinel-1 (S1) imagery (Fig. 1), with a range of 212 to 6591 individual samples per class. All spectral bands and polarizations from the three sensor sources have been included. The patch sizes were harmonized to the same extent of 60 × 60 m. The S1 and S2 scenes were chosen as closely as possible to the season and years of the aerial imagery which was taken around August between the years 2012 and 2020. For a better reproducibility of our experiments, we created a fixed split of train (90 %) and test (10 %) data. Detailed information about the datasets' structure and version history can be found in the description of Schulz et al. (2022). The data pre-processing and label derivation are further described in Sect. 2.3 and 2.4.
In this paper, we propose guidelines for the application of the latest machine learning techniques in forest remote sensing while handling an unweighted (i.e., real-world) dataset. In order to test the applicability of the multi-sensor imagery for species classification, we provide various benchmark experiments showcasing the information which can be extracted via various DL models. Specifically, we investigate the use of each sensor source individually, and multi-sensor fusion models examine the applicability of all imagery types present within the TreeSatAI Benchmark Archive.
The decision on the integration of multi-sensor data follows the state of research, which has shown that a large range of spectral information can be important for classifying species at a finer level based on its wavelengths. This could provide important discriminating information for classifying species with similar spatial and spectral features within the high-resolution RGB + NIR aerial data.
2.1 Study area
The TreeSatAI Benchmark Archive contains image data covering publicly managed forests in the federal state of Lower Saxony, Germany. The study area, covering approximately 47 710 km2, comprises both flat lands with a maritime climate (i.e., wet and winter mild) in the northwest as well as low mountain ranges with a more continental climate (i.e., dry and winter cold) in the southeast (Beck et al., 2018). Almost 22 % of the federal state consists of forests, of which large proportions belong to the continental parts of the study area (Holzwarth et al., 2020). According to NW-FVA (2021), the five predominant tree species in our research area were pine (Pinus sylvestris) with 38 %, beech (Fagus sylvatica) with 16 %, spruce (Picea abies) with 13 %, and oak (Quercus robur, Q. petraea) with 7 % of the forests. These species are usually grown as pure stands. The remaining tree stands cover 26 % of the research area. Among many others, they contain species such as ash (Fraxinus excelsior), Douglas fir (Pseudotsuga menziesii), larches (Larix decidua, L. kaempferi) and cherry (Prunus avium).
2.2 Reference data collection
For label extraction, we used forest administration data provided by the Lower Saxony State Forest Management Organisation (NLF, 2021a, b) (Fig. 2). The rolling archive is annually updated through spatially rotating terrestrial surveys and flight campaigns (Böckmann, 2016). It contains ground truth data and aerial imagery from 2011 to 2020, both corresponding to the same years.
Two forest administration datasets were chosen for label extraction. First of all, from the federal state forest management data (Waldeinrichtungsflächen der Niedersächsischen Landesforsten, WEFL), a subset of 175 142 vector geometries, covering approximately 318 000 ha of forested area, was made available. The dataset is derived from administrative surveys and interpretation of aerial imagery and collects information on the stand type and age class for single forest stands. Second, from the federal state forest inventory (Betriebsinventur der Niedersächsischen Landesforsten, BI), a subset of 18 289 point geometries was made available. The dataset covers permanent ground truth points with a 100 m distance grid over the study area (Saborowski et al., 2010). For a 13 m radius around each point, the BI collects a lot of information on the tree stand, including the main tree species, stand type, and stand age.
With the stand type, both datasets share a common attribute for derivation of tree species labels. However, both datasets have strengths and weaknesses in labeling quality and potential sample numbers. While having a very precise annotation from ground truth, the number of point features is limited in the BI data. In comparison, the high number of geometry features in the WEFL data allows for generation of large training datasets. Additionally, larger geometries can be used for the extraction of multiple samples for underrepresented classes. However, missing ground checks can in some cases lead to outdated or mislabeled information.
2.3 Remote sensing data collections
2.3.1 Aerial image collection
Aerial imagery from multiple flight campaigns from midsummer was provided with the forest administration data. The image collection covers four spectral bands (RGB + NIR) and has a spatial resolution of 0.2 m. With all the flight campaigns from the years 2011 to 2020, a full coverage of the study area is reached. The aerial imagery and the forest administration data have the same acquisition year and thus correspond to each other.
2.3.2 Sentinel-1 image collection
Especially by linking the backscatter information to tree crown volume and density (Cherrington et al., 2019), synthetic aperture radar (SAR) data could potentially be useful for tree species classification. Thus, we collected S1 scenes from mid-August over Lower Saxony for the years 2015 to 2020. The C-Band SAR sensor has a resolution of 10 m and a revisit time of less than 6 d (Torres et al., 2012). The S1 scenes were pre-processed into three-channel georeferenced image mosaics. The resulting images consist of VV and VH polarized channels and their ratio in the third channel.
As input for the processing pipeline, we used level-1 ground range detection (GRD) products in interferometric wide swath (IW) mode which were provided by the European Space Agency (ESA). With the S1TBX – ESA Sentinel-1 Toolbox from the Sentinel Application Platform (SNAP), we performed the following processing steps. We started with the application of orbit files, followed by border noise removal and radiometric calibration. For the orthorectification of the images, we applied range Doppler terrain correction with the digital elevation models SRTM30 for scenes below 60∘ and the ASTER GDEM otherwise. We converted the backscatter values from linear units to the decibel scale. As the last step, we updated the resolution of the imagery to 10 m using bicubic interpolation to match the spatial resolution of the S2 image collection.
2.3.3 Sentinel-2 image collection
To correspond to the other image collections temporally, we collected the S2 scenes over Lower Saxony for the years 2015 to 2020. The multi-spectral imagery of S2 contains surface and atmospheric reflectance values, including 13 bands in the visible and infrared ranges with a spatial resolution of 10 to 60 m (Drusch et al., 2012).
We atmospherically corrected all S2 Level-L1C scenes using SNAP-supported plugin Sen2Cor_v2.9 and resampled all bands to 10 m resolution. Both the data and software are provided by the ESA. The cirrus band was omitted due to a lack of information for most land surfaces. Although S2 imagery is collected by two satellites with a repetition rate between 3 and 5 d, multiple scenes from the same month can be affected by clouds. Additionally, for the years 2015 and 2016, only one of the two S2 satellites was in orbit. In order to cope with having observations from only one satellite and high cloud coverage, we computed cloud-free mosaics for the summer of each year, with median filtering over scenes from the months July, August, and September.
2.4 Sampling strategy
The goal of the sampling strategy was to generate highly reliable labeled image patches from all three image collections. This means that the timestamp of all the patches must fit the respective label year. For labels older than 2014 – when there were no S1 and S2 data available – we generated samples from the 2015 data to ensure annotations with the shortest temporal distance. To ensure the maximum number of image patches, we took both datasets WEFL and BI for label extraction.
Then, multiple sampling steps were done to derive single-labeled image patches (Fig. 3). (1) Class selection and aggregation. Based on the attribute stand type with 64 classes, we selected all classes with pure tree stands. If a pure class was not available for a tree species, we chose the respective pure and mixed or mixed classes and aggregated them. All other classes were excluded. After this step, 20 classes remained, which were re-labeled by the scientific name of the main tree species (Fig. 4). (2) Polygon sample selection (only WEFL). We excluded all polygons smaller than 2000 m2 to avoid the generation of image patches from very small tree stands. (3) Sample point creation (only WEFL). Within the polygons of each class, we created 5000 random points. They are supposed to be the center of the image patches. If possible, a negative buffer of 20 m to the polygon boundary was set to reduce the proximity of the points to neighboring tree stands. (4) Exclude points by distance threshold (only WEFL). To reduce the overlap of image patches, we excluded sample points with less than 20 m distance from each other. (5) Bounding box creation. For each of the sample points, a bounding box was created which depicts the size of the image patches. We chose 60 × 60 m bounding boxes leading to 300 × 300 pixels patches for the aerial imagery and 6 × 6 pixels patches for the S1 and S2 imagery. (6) Image patch generation. From the three image collections, we extracted the image patches according to the bounding boxes. Each file was named according to its genus, the species, the age class, the unique sample ID, the dataset, and the dataset source (e.g., Fagus_sylvatica_6_154201_BI_NLF.tif). For alternative labeling, we refer to the additional classes in Fig. 4. For multi-labeling files, we refer to the description of the Zenodo archive. (7) Sample number harmonization. Because of data gaps in the image collections, mainly caused by clouds in the S2 imagery, a large proportion of incomplete image triplets had to be removed. (8) Manual label check. From the remaining image patches, a minor number was wrongly labeled (e.g., different species) or had other issues (i.e., missing bands, haziness, blurriness). These images were manually removed through visual check-ups.
In order to provide some initial benchmarking of the presented dataset and to answer fundamental questions in the context of forest remote sensing, we evaluate and compare recent machine learning methods for uni-sensor and multi-sensor scene classification. We primarily focus on DL models given the main aim of this large-scale dataset is to enable the application of those methods for tree species classification. However, we also investigate the Light Gradient Boosting Machine (LightGBM) (Ke et al., 2017) in order to provide a comparison of tree-based machine learning methods with DL.
3.1 Deep learning models
For the aerial data, we apply ResNet18 (He et al., 2016) in order to model the spatial and spectral content. This model has shown high performance across various image processing tasks such as image classification (Scott et al., 2017), semantic segmentation (Chen et al., 2018), and content-based image retrieval (Sumbul et al., 2021). We assess training the network from scratch versus fine-tuning a pre-trained network. For the pre-trained network, we use model parameters obtained from training on ImageNet. Given that the ImageNet pre-trained models expect three bands as input, we also evaluate different configurations: (1) using only three of the possible four aerial bands and (2) duplicating one of the input channels in the pre-trained network in order to allow the use of all four aerial bands as input.
To model S1 and S2 data, we examine three different model architectures (Fig. 5). Given that the image patches are only 6 × 6 pixels in dimension, a standard 2-D convolutional network makes little sense as the spatial area to convolve is small. As such, we investigate three models which focus more on spectral rather than spatial content. The first model is a simple fully connected network in which each pixel in the S1 and S2 data is fully connected to each hidden node via a linear layer (Fig. 5a). The second model is a residual neural network (ResNet) utilizing 1-D convolutions instead of the standard 2-D. Convolutional neural networks (CNNs) using 1-D convolutions are often applied in hyperspectral classification problems (Ansari et al., 2021) where modeling spectral content, rather than spatial content, is the aim. We then test a 1-D ResNet (Hannun et al., 2019) for modeling the S1/S2 data (Fig. 5b). The final model is a vision transformer (ViT) (Dosovitskiy et al., 2020). The ViT performs image classification by breaking an image down into small patches which are used as input tokens. The information within each token, as well as the relationships between the tokens, is learned by the model. Given the small sizes, our Sentinel images need not be split into patches, and thus we use each individual band of the image as an input token. In this way, the model should learn the information within each band while simultaneously learning the relationships between the bands as well (Fig. 5c).
To jointly model S1 and S2 data with the aerial data, some form of fusion technique is required in order to combine the multi-sensor data. A naive approach is to simply concatenate the data along the channel dimension and then train a classifier (e.g., CNN). Such an approach is known as early fusion, given that the data are fused at the earliest possible stage. However, given the large discrepancies in spatial resolution between the aerial and S1/S2 data, concatenation at the image level is not possible without applying large amounts of interpolation to the Sentinel data. Late fusion is another technique which has shown greater performance than early fusion (Hong et al., 2020). Here, each modality is passed through its own separate network, and the fusion is performed on the final features. Such an approach allows for differences in the sizes of images from the various modalities since fusion is done once a single feature vector for each modality has been extracted. Thus, we focus our multi-sensor learning experiments on late fusion. In these experiments, the aerial data are modeled using ResNet, whereas the S1 and S2 branches each use one of the models mentioned above (Fig. 5).
3.2 Tree-based machine learning models
Another aspect we explore is the benefit of DL methods over traditional machine learning approaches. To this end, we use LightGBM (Ke et al., 2017), a framework for machine learning comprising several decision tree algorithms, including classification. Training DL models is energy- and time-consuming. In contrast, LightGBM provides algorithms for classification that are fast, memory efficient, and able to be executed in parallel. In other words, the cost of using a LightGBM classifier is significantly lower than that of training a deep model. Consequently, it is relevant to measure the effective accuracy boost of deep models in comparison to standard machine learning classifiers.
LightGBM classifiers were designed to solve single-label problems. Thus, since the TreeSatAI Benchmark Archive is a multi-label dataset, we cannot directly use LightGBM algorithms for tree species classification. To circumvent this, we employ a heuristic method known as one versus rest (OvR). In our scenario, OvR is used to cast a multi-label classification problem into multiple binary classification problems, one for each class.
For the classification of aerial images with LightGBM, we first use a ResNet18 pre-trained on ImageNet to extract features and then use these features to train the decision tree classifier. Here, we ignore the NIR band. Regarding fusion of aerial with S1 and/or S2, we concatenate the features predicted by the ResNet18 on the aerial imagery with data from S1, S2 or both and then train the LightGBM classifier.
In our experiments we aim to answer five main questions with respect to the dataset. The first question examines which of the three sensors is most suitable for the classification, as we have aerial data with a high spatial resolution and Sentinel data with a high range of spectral information. The second question is how to best train the model for aerial data. Here, we test three strategies for training our ResNet model: (1) using a model pre-trained on ImageNet data and retraining only the final classifier layer; (2) fine-tuning the weights of the final layer and classifier of a ResNet pre-trained on ImageNet; (3) training the model from scratch. The third question is whether the inclusion of the NIR band brings a significant benefit to the model performance. This is examined for two cases: (1) training from scratch; (2) fine-tuning a pre-trained model. Given that the model pre-trained on ImageNet typically only allows three bands as input, we duplicate the third input channel's weights when including the NIR band. The fourth question is to examine whether the fusion of aerial and Sentinel data improves the model performance compared to using only aerial data. As mentioned above, we use three different models in a late fusion setting for this. The final question we investigate is the ability of the various models to classify underrepresented classes within the dataset.
For all experiments, we trained models for a total of 150 epochs using a batch size of 32. For the learning rate, we use a cyclic learning rate scheduler (Smith, 2017) which modulates the learning rate value between 0.00005 and 0.001. In order to quantify model performance, we used recall, precision, F1 (Goutte and Gaussier, 2005), and mean average precision (mAP) (Everingham et al., 2010). To achieve a single summary statistic across the multiple classes being predicted, an averaging technique must be applied to the aforementioned metric scores. We chose to apply both micro and weighted averaging. Micro takes the global average, which shows the performance of the model without taking class imbalance into consideration. In order to also evaluate model performance with respect to class frequencies, the weighted average was included. Here, a given metric is calculated for each class, and the weighted average of the class scores (with the weights being the total number of instances for each label) is calculated (Takahashi et al., 2022).
4.1 Results of using individual input sources
Here, we examine the three data modalities of aerial, S1 and S2 available in the TreeSatAI Benchmark Archive. Table 1 shows that from the ResNet and multi-layer perceptron (MLP) models the best performance is achieved when using the aerial data, followed by S2 and then S1. In terms of weighted F1 and mAP, the models trained using aerial data outperformed those trained with S2 by 17 % and 13 %, respectively. For the micro F1 and mAP, using aerial data outperforms S2 by 15 % and 12 %, respectively. This indicates that the spatial content present within the aerial data is very informative for the classification of the various tree species. The lower S2 and S1 scores are likely because of the lower spatial resolution of 10 m compared to aerial with 0.2 m per pixel. However, the larger number of spectral bands in S2 still allows for mAP scores of about 65 % and F1 of around 53 %. When examining the scores obtained using S1 data, we can see that using the 10 m resolution back-scattering information alone may not be sufficient for tree species classification.
The LightGBM model results show similar scores for S1 and S2. However, for the aerial data the model reaches lower weighted F1 and mAP scores with losses of 25 % and 17 %, respectively. This indicates that extracting features with a ResNet18 and using these features to train a decision tree classifier is not recommended for aerial images.
4.2 Analysis of different training strategies for aerial data
In this section, we discuss whether using a pre-trained network performs better than training the network from scratch. We thus test the same three training scenarios as in the previous subsection, but here we focus only on comparing the best model from each scenario. In Table 2, we see that the Fine-tuning All network leads to slightly improved metric performance when compared to Scratch, while the Retrain Head scenario performed the worst. Comparing Fine-tuning All with Scratch, we see roughly 1 % performance increases for both weighted F1 and weighted mAP, while the micro average of F1 was equal. Comparing these two models further, we see that the micro mAP was higher by 1 % for Scratch as well as the weighted and micro precisions, which were higher than the Fine-tuning All by 1 % and 3 %, respectively. This suggests that training from scratch can achieve a nearly identical performance to using a pre-trained network. The Retrain Head variant showed a very low performance, achieving roughly half the performance as the Fine-tuning All variant for both F1 and mAP scores. This indicates that the features learned from training on ImageNet need fine-tuning when applied to remote sensing image classification tasks. Given that the pre-trained network does slightly outperform the network trained from scratch, we apply the three-band fine-tuned network for the aerial data in all multi-sensor learning approaches.
4.3 Effect of the near-infrared band
In this section, we examine the influence of including the NIR band when training the network on the aerial data. We tested this across three training scenarios: (1) training the model from scratch (Scratch); (2) fine-tuning all layers of a pre-trained network (Fine-tuning All); (3) fixing all the convolutional layer weights from a pre-trained network and only training the classification head of the model (Retrain Head). Pre-trained models used weights obtained by training on ImageNet.
As seen in Table 2, the NIR band does help when we train from scratch, as both the micro and weighted F1 and mAP scores all increased by 2 % compared to using only RGB inputs. This is to be expected as the NIR band is an important indicator of vegetation monitoring (Tucker, 1979).
When we look at both variations of the pre-trained models, the result is the opposite, as the F1 and mAP scores are higher when using only the RGB channels. This is likely due to both pre-trained models having been pre-trained on RGB inputs. As such, the duplicated channel used for NIR in the pre-trained scenarios came from a channel which was not originally optimized for NIR input. Given that the value ranges of the NIR channel are higher than those of any of the RGB channels, the additional duplicated channel negatively influenced the pre-trained model performances. The three-band Fine-tuning All variant obtained 1 % metric improvements over the four-band variant in F1 and mAP. Likewise, the three-band Retrain Head outperformed the four-band variant by 6 %–11 % in the F1 and mAP metrics. For the four-band Retrain Head, the duplicated channel was not retrained for the NIR input, and thus the model was unable to adapt the network weights for the NIR input, leading to larger discrepancies between the three- and four-band performances. The Fine-tune All model was better able to adapt the duplicated channel to the NIR input as its pre-trained weights were further trained.
4.4 Results of multi-sensor fusion
Although the earlier results from Table 1 indicate that the information from S1 and S2 is weaker than that of the aerial data, the combination of the various modalities may provide a greater benefit than their individual use. Thus, we examine the performance of various multi-sensor learning techniques. Table 3 shows the results for different models applied to the multi-sensor fusion of aerial and Sentinel data. We see that, in terms of F1 and mAP scores, simple MLP performs best, outperforming the 1-D CNN and Transformer variants by about 1 % across both micro and weighted F1 and mAP. This slight increase in metric performance may be due to the full connections within the MLP network being better suited to modeling all of the information present within the small Sentinel patches.
All DL-based methods outperformed the LightGBM baseline approach in terms of F1 and mAP scores. However, LightGBM did manage to obtain the highest weighted precision score using the combination of aerial and S2 and the highest micro precision using the combination of aerial, S1 and S2. On the other hand, the mAP scores for LightGBM were about 10 % lower than the other fusion models, indicating that LightGBM performed well at the 50 % threshold for a positive prediction but had lower performance across the various thresholds calculated in the mAP.
When compared to the uni-sensor approach which only utilizes aerial data, the incorporation of S1 and S2 data provides an increase of 1 % for the weighted F1 and mAP and a 2 % increase in the micro F1 and mAP. Thus, the high-resolution aerial imagery provides the most important features for differentiating the various species, while the spatial resolution of the S1 and S2 data is likely too low to provide a larger benefit.
4.5 Learning underrepresented classes
Here, we examine the class-based performances using the best-performing model (multi-sensor late fusion using MLP, Table 4). The model performance is highest for classes Pinus (82 % F1, 73 % mAP), Picea (79 % F1, 69 % mAP), Abies (77 % F1, 61 % mAP) and Quercus (71 % F1, 60 % mAP). The model performed worst when predicting Tilia (47 % F1, 23 % mAP), Prunus (52 % F1, 31 % mAP), Acer (53 % F1, 31 % mAP), Betula (54 % F1, 35 % mAP) and Fraxinus (58 % F1, 36 % mAP). The majority of the classes for which the model was most successful corresponded to the most frequently occurring classes, whereas the classes for which the model performed poorly were those with few image samples. This class imbalance mirrors the real-world situation at our study site. The one exception was the class Abies, which made up only 2 % of the image samples but which had the third-highest classification scores. This result suggests that the tree stands with silver fir (Abies alba) contain some easily identifiable features that set it apart from other classes.
The TreeSatAI Benchmark Archive was made available through Zenodo (https://doi.org/10.5281/zenodo.6598390, Schulz et al., 2022) under Creative Commons Attribution 4.0 International. Full code examples are published on the GitHub repositories of the Remote Sensing Image Analysis (RSiM) group (https://git.tu-berlin.de/rsim/treesat_benchmark, last access: 2 February 2023; Schulz et al., 2022) and the Deutsches Forschungszentrum für Künstliche Intelligenz (DFKI) (https://github.com/DFKI/treesatai_benchmark, last access: 2 February 2023; Schulz et al., 2022). Code examples for the sampling strategy can be made available by the corresponding author via email request.
In this paper, we introduced a novel dataset for multi-label and single-label tree species classification from aerial, S1 and S2 imagery. It includes more than 50 000 temporally (i.e., same season and same year) and spatially harmonized image patch triplets from the years 2011 to 2020 with a size of 60 × 60 m from all three sensors. The labels for 20 tree species (i.e., 15 tree genera) were derived from forest administration data from northern Germany.
Several DL and tree-based machine learning models using the dataset have been evaluated. With regards to the benchmarking results, we find that a pre-trained DL network using only the RGB bands of the 0.2 m resolution aerial data performs best for modeling. The incorporation of additional spectral information from the S1 and S2 sensors led to a small increase in performance. We found that the utilization of airborne data performs sufficiently well for the study task, suggesting that high-resolution RGB imagery provides enough information for DL models to differentiate forest stands at the tree species level.
DL models only using aerial imagery reached high precisions (>80 %) for 6 of the 15 classes. This result demonstrates the potential to map predominant tree species at the 0.2 m resolution (cf. Ahlswede et al., 2022). However, our class-wise model performances using mono-temporal (i.e., mono-seasonal summer scenes from different years) data could not reach the accuracies of previous studies exploiting S2 multi-temporal (i.e., multi-seasonal) data (e.g., Immitzer et al., 2019; Grabska et al., 2019). Nonetheless, we expect that our model approaches using multi-sensor fusion are robust to new data from different spatial areas.
Similarly to many forest remote sensing studies, our dataset contains a strong class imbalance which negatively affected the performance of underrepresented classes. Future works based on the existing dataset could thus examine new methods for dealing with data imbalance. The simple weighting of the loss function based on class frequencies, which we applied, was not sufficient to overcome this issue. In addition, further studies can focus on improving the exploitation of the S2 multi-spectral data, as the models tested in this work were only capable of minor improvements over the RGB + NIR aerial imagery data.
For further improvements of the TreeSatAI Benchmark Archive, additional remote sensing products or label samples could be integrated. First, the integration of multi-seasonal data might disentangle further species-related information regarding phenology phases (e.g., spring flush, second flush, or autumnal leaf fall) or different branch structure types during the dormant phase (Kowalski et al., 2020; Kollert et al., 2021). This information could be derived from S2 and S1 image time series and additional aerial imagery from the winter season. Second, active data with a higher spatial resolution (e.g., lidar) may improve the characterization of species-related tree crown features. Third, higher spectral resolution (i.e., hyperspectral) data, e.g., from the Environmental Mapping and Analysis Programme (EnMAP), may also reveal biophysical components differentiating tree species. Even though most of the aforementioned remote sensing products are not available for the same time frame, a combination could still be beneficial for tree species classification (Fricker et al., 2019; Weinstein et al., 2021; Mäyrä et al., 2021).
We highly encourage potential users of the TreeSatAI Benchmark Archive to create additional layers, generate further samples from other spatial domains or develop new model architectures. This benchmark archive and its future applications can support the provision of large-scale tree species maps, helping forest authorities to face challenges driven by global environmental changes.
All the authors were involved in the conceptualization and methodology of the paper. Data investigation, compilation and dataset preparation were done by CS, SA, PH, BB, FA, JH, and BK. The analysis and code development were done by SA, CG, and JH. SA and CS wrote the original draft under the supervision of BD, BK and MF. All the authors were involved in the review and editing process. The TreeSatAI project was administrated by BK.
The contact author has declared that none of the authors has any competing interests.
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
We are grateful to the Niedersächsische Landesforsten for providing us with the aerial imagery and the reference data. We thank the European Space Agency for providing us with the pre-processed Sentinel-1 and Sentinel-2 data and the respective software tools. Stenka Vulova kindly improved our English.
TreeSatAI was funded by the German Federal Ministry of Education and Research under grant no. 01IS20014A.
This paper was edited by Jia Yang and reviewed by two anonymous referees.
Ahlswede, S., Thekke-Madam, N., Schulz, C., Kleinschmit, B., and Demir, B.: Weakly Supervised Semantic Segmentation of Remote Sensing Images for Tree Species Classification Based on Explanation Methods, in: IEEE International Geoscience and Remote Sensing Symposium, 17–22 July 2022, Kuala Lumpur, Malaysia, https://doi.org/10.48550/arXiv.2201.07495, 2022. a
Ansari, M., Homayouni, S., Safari, A., and Niazmardi, S.: A New Convolutional Kernel Classifier for Hyperspectral Image Classification, IEEE J. Sel. Top. Appl., 14, 11240–11256, 2021. a
Basu, S., Ganguly, S., Mukhopadhyay, S., DiBiano, R., Karki, M., and Nemani, R.: Deepsat: a learning framework for satellite imagery, in: Proceedings of the 23rd SIGSPATIAL international conference on advances in geographic information systems, 3–6 November 2015, Seattle, Washington, USA, 1–10, https://doi.org/10.48550/arXiv.1509.03602, 2015. a
Beck, H. E., Zimmermann, N. E., McVicar, T. R., Vergopolan, N., Berg, A., and Wood, E. F.: Present and future Köppen-Geiger climate classification maps at 1-km resolution, Scientific Data, 5, 1–12, 2018. a
Böckmann, T.: Warum sind Betriebsinventuren für die forstliche Praxis wichtig (Why is two-phase sampling for stratification so important for forestry enterprises?), Forstarchiv, 87, 31–37, 2016. a
Boser, B. E., Guyon, I. M., and Vapnik, V. N.: A training algorithm for optimal margin classifiers, in: Proceedings of the fifth annual workshop on Computational learning theory, 27–29 July 1992, New York, United States, 144–152, https://doi.org/10.1145/130385.130401, 1992. a
Breiman, L.: Random forests, Mach. Learn., 45, 5–32, 2001. a
Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation, in: Proceedings of the European conference on computer vision (ECCV), 8–14 September 2018, Munich, Germany, 801–818, https://doi.org/10.48550/arXiv.1802.02611, 2018. a
Cherrington, E., Flores-Anderson, A., Thapa, R. B., Herndon, K. E., Wahome, A., Oduor, P., Mubea, K., Ouko, E., and Hanh, N.: Perspectives on the Future Application of SAR in Forest and Environmental Monitoring, in: The SAR handbook: comprehensive methodologies for forest monitoring and biomass estimation, edited by: Flores-Anderson, A. I., Herndon, K. E., Thapa, R. B., and Cherrington, E., chap. 8, NASA, https://doi.org/10.25966/nr2c-s697, 2019. a
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L.: Imagenet: A large-scale hierarchical image database, in: 2009 IEEE conference on computer vision and pattern recognition, IEEE, 20–25 June 2009, Miami, Florida, USA, 248–255, https://doi.org/10.1109/CVPR.2009.5206848, 2009. a
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., and Houlsby, N.: An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale, in: International Conference on Learning Representations, 26 April–1 May 2020, Addis Ababa, Ethiopia, https://doi.org/10.48550/arXiv.2010.11929, 2020. a
Drusch, M., Del Bello, U., Carlier, S., Colin, O., Fernandez, V., Gascon, F., Hoersch, B., Isola, C., Laberinti, P., Martimort, P., Meygret, A., Spoto, F., Sy, O., Marchese, F., and Bargellini, P.: Sentinel-2: ESA's optical high-resolution mission for GMES operational services, Remote Sens. Environ., 120, 25–36, 2012. a
Egli, S. and Höpke, M.: CNN-Based Tree Species Classification Using High Resolution RGB Image Data from Automated UAV Observations, Remote Sens., 12, 3892, https://doi.org/10.3390/rs12233892, 2020. a
Everingham, M., Van Gool, L., Williams, C. K., Winn, J., and Zisserman, A.: The pascal visual object classes (voc) challenge, Int. J. Comput. Vision, 88, 303–338, 2010. a
Fassnacht, F. E., Latifi, H., Stereńczak, K., Modzelewska, A., Lefsky, M., Waser, L. T., Straub, C., and Ghosh, A.: Review of studies on tree species classification from remotely sensed data, Remote Sens. Environ., 186, 64–87, 2016. a
Fricker, G. A., Ventura, J. D., Wolf, J. A., North, M. P., Davis, F. W., and Franklin, J.: A Convolutional Neural Network Classifier Identifies Tree Species in Mixed-Conifer Forest from Hyperspectral Imagery, Remote Sensing, 11, 2326, https://doi.org/10.3390/rs11192326, 2019. a
Ganz, S., Adler, P., and Kändler, G.: Forest Cover Mapping Based on a Combination of Aerial Images and Sentinel-2 Satellite Data Compared to National Forest Inventory Data, Forests, 11, 1322, https://doi.org/10.3390/f11121322, 2020. a, b
Goutte, C. and Gaussier, E.: A probabilistic interpretation of precision, recall and F-score, with implication for evaluation, in: Advances in Information Retrieval. Lecture Notes in Computer Science, Vol. 3408, edited by: Losada, D. E. and Fernández-Luna, J. M., Springer, 345–359, https://doi.org/10.1007/978-3-540-31865-1_25, 2005. a
Grabska, E., Hostert, P., Pflugmacher, D., and Ostapowicz, K.: Forest stand species mapping using the Sentinel-2 time series, Remote Sens., 11, 1197, https://doi.org/10.3390/rs11101197, 2019. a, b
Hamedianfar, A., Mohamedou, C., Kangas, A., and Vauhkonen, J.: Deep learning for forest inventory and planning: a critical review on the remote sensing approaches so far and prospects for further applications, Forestry, 95, 451–465, 2022. a, b
Hannun, A. Y., Rajpurkar, P., Haghpanahi, M., Tison, G. H., Bourn, C., Turakhia, M. P., and Ng, A. Y.: Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network, Nat. Med., 25, 65–69, 2019. a
He, K., Zhang, X., Ren, S., and Sun, J.: Deep residual learning for image recognition, in: Proceedings of the IEEE conference on computer vision and pattern recognition, 27–30 June 2016, Las Vegas, Nevada, USA, 770–778, https://doi.org/10.1109/CVPR.2016.90, 2016. a
Helber, P., Bischke, B., Dengel, A., and Borth, D.: EuroSAT: A Novel Dataset and Deep Learning Benchmark for Land Use and Land Cover Classification, IEEE J. Sel. Top. Appl., 12, 2217–2226, 2019. a
Hemmerling, J., Pflugmacher, D., and Hostert, P.: Mapping temperate forest tree species using dense Sentinel-2 time series, Remote Sens. Environ., 267, 112743, https://doi.org/10.1016/j.rse.2021.112743, 2021. a
Hlásny, T., Barka, I., Roessiger, J., Kulla, L., Trombik, J., Sarvašová, Z., Bucha, T., Kovalčík, M., and Čihák, T.: Conversion of Norway spruce forests in the face of climate change: a case study in Central Europe, Eur. J. For. Res., 136, 1013–1028, 2017. a
Hollaus, M. and Vreugdenhil, M.: Radar satellite imagery for detecting bark beetle outbreaks in forests, Current Forestry Reports, 5, 240–250, 2019. a
Holzwarth, S., Thonfeld, F., Abdullahi, S., Asam, S., Da Ponte Canova, E., Gessner, U., Huth, J., Kraus, T., Leutner, B., and Kuenzer, C.: Earth observation based monitoring of forests in Germany: A review, Remote Sens., 12, 3570, https://doi.org/10.3390/rs12213570, 2020. a, b, c, d
Hong, D., Gao, L., Yokoya, N., Yao, J., Chanussot, J., Du, Q., and Zhang, B.: More diverse means better: Multimodal deep learning meets remote-sensing imagery classification, IEEE T. Geosci. Remote Sens., 59, 4340–4354, 2020. a
Immitzer, M., Neuwirth, M., Böck, S., Brenner, H., Vuolo, F., and Atzberger, C.: Optimal input features for tree species classification in Central Europe based on multi-temporal Sentinel-2 data, Remote Sens., 11, 2599, https://doi.org/10.3390/rs11222599, 2019. a, b
IPCC: Climate Change 2014: Synthesis Report. Contribution of Working Groups I, II and III to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change, IPCC, Geneva, Switzerland, https://www.ipcc.ch/report/ar5/syr/ (last access: 31 January 2023), 2014. a
Karlson, M., Ostwald, M., Reese, H., Sanou, J., Tankoano, B., and Mattsson, E.: Mapping tree canopy cover and aboveground biomass in Sudano-Sahelian woodlands using Landsat 8 and random forest, Remote Sens., 7, 10017–10041, 2015. a
Kattenborn, T., Eichel, J., Wiser, S., Burrows, L., Fassnacht, F. E., and Schmidtlein, S.: Convolutional Neural Networks accurately predict cover fractions of plant species and communities in Unmanned Aerial Vehicle imagery, Remote Sensing in Ecology and Conservation, 6, 472–486, 2020. a
Kattenborn, T., Leitloff, J., Schiefer, F., and Hinz, S.: Review on Convolutional Neural Networks (CNN) in vegetation remote sensing, ISPRS J. Photogramm., 173, 24–49, 2021. a
Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., Ye, Q., and Liu, T.-Y.: LightGBM: A Highly Efficient Gradient Boosting Decision Tree, Adv. Neur. In., 30, 3146–3154, 2017. a, b
Kollert, A., Bremer, M., Löw, M., and Rutzinger, M.: Exploring the potential of land surface phenology and seasonal cloud free composites of one year of Sentinel-2 imagery for tree species mapping in a mountainous region, Int. J. Appl. Earth Obs., 94, 102208, 2021. a, b
Kowalski, K., Senf, C., Hostert, P., and Pflugmacher, D.: Characterizing spring phenology of temperate broadleaf forests using Landsat and Sentinel-2 time series, Int. J. Appl. Earth Obs., 92, 102172, https://doi.org/10.1016/j.jag.2020.102172, 2020. a, b
Krzystek, P., Serebryanyk, A., Schnörr, C., Červenka, J., and Heurich, M.: Large-Scale Mapping of Tree Species and Dead Trees in Šumava National Park and Bavarian Forest National Park Using Lidar and Multispectral Imagery, Remote Sensing, 12, 661, https://doi.org/10.3390/rs12040661, 2020. a
LGLN: Geodatenportal Niedersachsen (GDI-NI), Landesamt für Geoinformation und Landesvermessung Niedersachsen, https://www.geodaten.niedersachsen.de/, last access: 31 December 2020. a
MacDicken, K., Jonsson, Ö., Piña, L., Maulo, S., Contessa, V., Adikari, Y., Garzuglia, M., Lindquist, E., Reams, G., and D'Annunzio, R.: Global forest resources assessment 2015: how are the world's forests changing?, FAO, https://agris.fao.org/agris-search/search.do?recordID=XF2017001127 (last access: 31 January 2023), 2016. a
Martins, G. B., La Rosa, L. E. C., Happ, P. N., Coelho Filho, L. C. T., Santos, C. J. F., Feitosa, R. Q., and Ferreira, M. P.: Deep learning-based tree species mapping in a highly diverse tropical urban setting, Urban For. Urban Gree., 64, 127241, https://doi.org/10.1016/j.ufug.2021.127241, 2021. a
Mäyrä, J., Keski-Saari, S., Kivinen, S., Tanhuanpää, T., Hurskainen, P., Kullberg, P., Poikolainen, L., Viinikka, A., Tuominen, S., Kumpula, T., and Vihervaara, P.: Tree species classification from airborne hyperspectral and LiDAR data using 3D convolutional neural networks, Remote Sens. Environ., 256, 112322, https://doi.org/10.1016/j.rse.2021.112322, 2021. a
NLF: Betriebsinventurdaten der Niedersächsischen Landesforsten (BI), Niedersächsische Landesforsten, 2021a. a, b, c, d
NLF: Waldeinrichtungsflächen der Niedersächsischen Landesforsten (WEFL), Niedersächsische Landesforsten (Lower Saxony State Forest Management Organisation), 2021b. a, b, c, d
NW-FVA: Waldzustandsbericht 2021 für Niedersachsen, Nordwestdeutsche Forstliche Versuchsanstalt, Niedersächsisches Ministerium für Ernährung, Landwirtschaft und Verbraucherschutz, Zenodo, https://doi.org/10.5281/zenodo.5615008, 2021. a
Ottosen, T.-B., Petch, G., Hanson, M., and Skjøth, C. A.: Tree cover mapping based on Sentinel-2 images demonstrate high thematic accuracy in Europe, Int. J. Appl. Earth Obs., 84, 101947, https://doi.org/10.1016/j.jag.2019.101947, 2020. a
Pasquarella, V. J., Holden, C. E., and Woodcock, C. E.: Improved mapping of forest type using spectral-temporal Landsat features, Remote Sens. Environ., 210, 193–207, 2018. a
Saborowski, J., Marx, A., Nagel, J., and Böckmann, T.: Double sampling for stratification in periodic inventories—Infinite population approach, Forest Ecol. Manag., 260, 1886–1895, 2010. a
Schiefer, F., Kattenborn, T., Frick, A., Frey, J., Schall, P., Koch, B., and Schmidtlein, S.: Mapping forest tree species in high resolution UAV-based RGB-imagery by means of convolutional neural networks, ISPRS J. Photogramm., 170, 205–215, 2020. a, b
Schmitt, M., Hughes, L. H., Qiu, C., and Zhu, X. X.: SEN12MS–A Curated Dataset of Georeferenced Multi-Spectral Sentinel-1/2 Imagery for Deep Learning and Data Fusion, arXiv [preprint], https://doi.org/10.48550/arXiv.1906.07789, 2019. a
Schuldt, B., Buras, A., Arend, M., Vitasse, Y., Beierkuhnlein, C., Damm, A., Gharun, M., Grams, T. E., Hauck, M., Hajek, P., Hartmann, H., Hiltbrunner, E., Hoch, G., Holloway-Phillips, M., Körner, C., Larysch, E., Lübbe, T., Nelson, D., Rammig, A., Rigling, A., Rose, L., Ruehr, N., Schumann, K., Weiser, F., Werner, C., Wohlgemuth, T., Zang, C., and Kahmen, A.: A first assessment of the impact of the extreme 2018 summer drought on Central European forests, Basic Appl. Ecol., 45, 86–103, 2020. a
Schulz, C., Ahlswede, S., Gava, C., Helber, P., Bischke, B., Arias, F., Förster, M., Hees, J., Demir, B., and Kleinschmit, B.: TreeSatAI Benchmark Archive for Deep Learning in Forest Applications (1.0.1), Zenodo [code, data set], https://doi.org/10.5281/zenodo.6598390, 2022. a, b, c, d, e, f
Scott, G. J., England, M. R., Starms, W. A., Marcum, R. A., and Davis, C. H.: Training deep convolutional neural networks for land–cover classification of high-resolution imagery, IEEE Geosci. Remote S., 14, 549–553, 2017. a
Senf, C., Buras, A., Zang, C. S., Rammig, A., and Seidl, R.: Excess forest mortality is consistently linked to drought across Europe, Nat. Commun., 11, 1–8, 2020. a
Sesnie, S. E., Finegan, B., Gessler, P. E., Thessler, S., Ramos Bendana, Z., and Smith, A. M.: The multispectral separability of Costa Rican rainforest types with support vector machines and Random Forest decision trees, Int. J. Remote Sens., 31, 2885–2909, 2010. a
Smith, L. N.: Cyclical learning rates for training neural networks, in: 2017 IEEE winter conference on applications of computer vision (WACV), IEEE, 24–31 March 2017, Santa Rosa, CA, USA, 464–472, https://doi.org/10.1109/WACV.2017.58, 2017. a
Sumbul, G., Charfuelan, M., Demir, B., and Markl, V.: Bigearthnet: A Large-Scale Benchmark Archive for Remote Sensing Image Understanding, in: IEEE International Geoscience and Remote Sensing Symposium, 28 July–2 August 2019, Yokohama, Japan, 5901–5904, https://doi.org/10.1109/IGARSS.2019.8900532, 2019. a, b
Sumbul, G., De Wall, A., Kreuziger, T., Marcelino, F., Costa, H., Benevides, P., Caetano, M., Demir, B., and Markl, V.: BigEarthNet-MM: A Large-Scale, Multimodal, Multilabel Benchmark Archive for Remote Sensing Image Classification and Retrieval, IEEE Geoscience and Remote Sensing Magazine, 9, 174–180, 2021. a
Takahashi, K., Yamamoto, K., Kuchiba, A., and Koyama, T.: Confidence interval for micro-averaged F1 and macro-averaged F1 scores, Appl. Intell., 52, 4961–4972, 2022. a
Tanase, M. A., Aponte, C., Mermoz, S., Bouvet, A., Le Toan, T., and Heurich, M.: Detection of windthrows and insect outbreaks by L-band SAR: A case study in the Bavarian Forest National Park, Remote Sens. Environ., 209, 700–711, 2018. a
Thonfeld, F., Gessner, U., Holzwarth, S., Kriese, J., da Ponte, E., Huth, J., and Kuenzer, C.: A First Assessment of Canopy Cover Loss in Germany's Forests after the 2018–2020 Drought Years, Remote Sens., 14, 562, https://doi.org/10.3390/rs14030562, 2022. a
Torres, R., Snoeij, P., Geudtner, D., Bibby, D., Davidson, M., Attema, E., Potin, P., Rommen, B., Floury, N., Brown, M., Traver, I. N., Deghaye, P., Duesmann, B., Rosich, B., Miranda, N., Bruno, C., L'Abbate, M., Croci, R., Pietropaolo, A., Huchler, M., and Rostan, F.: GMES Sentinel-1 mission, Remote Sens. Environ., 120, 9–24, 2012. a
Tucker, C. J.: Red and photographic infrared linear combinations for monitoring vegetation, Remote Sens. Environ., 8, 127–150, 1979. a
Waser, L. T., Rüetschi, M., Psomas, A., Small, D., and Rehush, N.: Mapping dominant leaf type based on combined Sentinel-1/-2 data–Challenges for mountainous countries, ISPRS J. Photogramm., 180, 209–226, 2021. a
Weinstein, B. G., Marconi, S., Bohlman, S., Zare, A., and White, E.: Individual tree-crown detection in RGB imagery using semi-supervised deep learning neural networks, Remote Sensing, 11, 1309, https://doi.org/10.3390/rs11111309, 2019a. a, b
Weinstein, B. G., Marconi, S., Bohlman, S. A., Zare, A., and White, E. P.: Geographic generalization in airborne RGB deep learning tree detection, bioRxiv [preprint], https://doi.org/10.1101/790071, 2019b. a
Weinstein, B. G., Marconi, S., Bohlman, S., Zare, A., Singh, A., Graves, S. J., and White, E.: A remote sensing derived data set of 100 million individual tree crowns for the National Ecological Observatory Network, eLife, 10, e62922, https://doi.org/10.7554/eLife.62922, 2021. a, b
Welle, T., Aschenbrenner, L., Kuonath, K., Kirmaier, S., and Franke, J.: Mapping Dominant Tree Species of German Forests, Remote Sens., 14, 3330, https://doi.org/10.3390/rs14143330, 2022. a
Wurm, M., Stark, T., Zhu, X. X., Weigand, M., and Taubenböck, H.: Semantic segmentation of slums in satellite images using transfer learning on fully convolutional neural networks, ISPRS J. Photogramm., 150, 59–69, 2019. a
Xu, K., Zhang, Z., Yu, W., Zhao, P., Yue, J., Deng, Y., and Geng, J.: How Spatial Resolution Affects Forest Phenology and Tree-Species Classification Based on Satellite and Up-Scaled Time-Series Images, Remote Sens., 13, 2716, https://doi.org/10.3390/rs13142716, 2021. a
Yang, Y. and Newsam, S.: Bag-Of-Visual-Words and Spatial Extensions for Land-Use Classification, ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, 2–5 November 2005, San Jose, California, USA, https://doi.org/10.1145/1869790.1869829, 2010. a
Zhang, C., Xia, K., Feng, H., Yang, Y., and Du, X.: Tree species classification using deep learning and RGB optical images obtained by an unmanned aerial vehicle, J. Forestry Res., 32, 1879–1888, 2021. a