Table of Contents
Fetching ...

Planted: a dataset for planted forest identification from multi-satellite time series

Luis Miguel Pazos-Outón, Cristina Nader Vasconcelos, Anton Raichuk, Anurag Arnab, Dan Morris, Maxim Neumann

TL;DR

The paper addresses global monitoring of planted forests and tree crops using multi-satellite time series. It introduces Planted, a large, curated multimodal dataset spanning five satellites with 2,264,747 examples across 64 classes and 41 countries, along with labels and metadata. The authors establish baseline single-modality and multimodal transformer baselines, analyze data augmentation and fusion strategies, and show mid-fusion across 2-3 modalities yields strong performance while highlighting challenges like missing data and label imbalance. The dataset provides a benchmark to spur research in multimodal, time-series remote sensing for forest monitoring with potential impact on conservation and carbon accounting.

Abstract

Protecting and restoring forest ecosystems is critical for biodiversity conservation and carbon sequestration. Forest monitoring on a global scale is essential for prioritizing and assessing conservation efforts. Satellite-based remote sensing is the only viable solution for providing global coverage, but to date, large-scale forest monitoring is limited to single modalities and single time points. In this paper, we present a dataset consisting of data from five public satellites for recognizing forest plantations and planted tree species across the globe. Each satellite modality consists of a multi-year time series. The dataset, named \PlantD, includes over 2M examples of 64 tree label classes (46 genera and 40 species), distributed among 41 countries. This dataset is released to foster research in forest monitoring using multimodal, multi-scale, multi-temporal data sources. Additionally, we present initial baseline results and evaluate modality fusion and data augmentation approaches for this dataset.

Planted: a dataset for planted forest identification from multi-satellite time series

TL;DR

The paper addresses global monitoring of planted forests and tree crops using multi-satellite time series. It introduces Planted, a large, curated multimodal dataset spanning five satellites with 2,264,747 examples across 64 classes and 41 countries, along with labels and metadata. The authors establish baseline single-modality and multimodal transformer baselines, analyze data augmentation and fusion strategies, and show mid-fusion across 2-3 modalities yields strong performance while highlighting challenges like missing data and label imbalance. The dataset provides a benchmark to spur research in multimodal, time-series remote sensing for forest monitoring with potential impact on conservation and carbon accounting.

Abstract

Protecting and restoring forest ecosystems is critical for biodiversity conservation and carbon sequestration. Forest monitoring on a global scale is essential for prioritizing and assessing conservation efforts. Satellite-based remote sensing is the only viable solution for providing global coverage, but to date, large-scale forest monitoring is limited to single modalities and single time points. In this paper, we present a dataset consisting of data from five public satellites for recognizing forest plantations and planted tree species across the globe. Each satellite modality consists of a multi-year time series. The dataset, named \PlantD, includes over 2M examples of 64 tree label classes (46 genera and 40 species), distributed among 41 countries. This dataset is released to foster research in forest monitoring using multimodal, multi-scale, multi-temporal data sources. Additionally, we present initial baseline results and evaluate modality fusion and data augmentation approaches for this dataset.

Paper Structure

This paper contains 31 sections, 2 equations, 7 figures, 15 tables.

Figures (7)

  • Figure 1: Overview of the Planted dataset and the geographical samples distribution.
  • Figure 2: Temporal signatures of randomly selected tree species across MODIS spectral reflectance bands and NDVI and EVI indicators.
  • Figure 3: Depiction of the multimodal transformer architecture. In early fusion all cross-modal layers are activated, in mid fusion the early layers are excluded and we rely on the mid layer and the final layers. In late fusion only the last cross-modal layers are used.
  • Figure 4: Patch size ablations for each satellite. The best patch size is satellite-dependent. The notation above refers to time x spatial dimensions.
  • Figure 5: Performance for different fusion layers. The green line shows the top performing combination of satellites, and the blue line the average performance across all combinations of satellites.
  • ...and 2 more figures