A curated list of open technology projects to sustain a stable climate, energy supply, biodiversity and natural resources.

A high-resolution canopy height model of the Earth

The model estimates canopy top height for every Sentinel-2 image pixel and was trained using sparse GEDI LIDAR data as a reference.
https://github.com/langnico/global-canopy-height-model

Category: Biosphere
Sub Category: Forest Modeling and Analysis

Keywords

canopy-height-models deep-learning ensemble gedi sentinel-2 uncertainty

Last synced: about 2 hours ago
JSON representation

Repository metadata

This repository contains the code used in the paper: A high-resolution canopy height model of the Earth. Here, we developed a model to estimate canopy top height anywhere on Earth. The model estimates canopy top height for every Sentinel-2 image pixel and was trained using sparse GEDI LIDAR data as a reference.

README.md

A high-resolution canopy height model of the Earth

This repository contains the code used to create the results presented in the paper: A high-resolution canopy height model of the Earth.
Here, we developed a model to estimate canopy top height anywhere on Earth. The model estimates canopy top height for every Sentinel-2 image pixel and was trained using sparse GEDI LIDAR data as a reference.

See our project page for an interactive demo and more information.

Table of Contents

  1. Data availability
  2. Installation and credentials
  3. Loading the model
  4. Deploying
  5. Training
  6. ALS preprocessing for independent comparison
  7. Citation

Data availability

This is a summary of all the published data:

Installation and credentials

Please follow the instructions in INSTALL.md.

Loading the model

from gchm.models.xception_sentinel2 import xceptionS2_08blocks_256
# load the model with random initialization
model = xceptionS2_08blocks_256()

Please see the example notebook on how to load the model with the trained weights.

Deploying

This is a demo how to run the trained ensemble to compute the canopy height map for a Sentinel-2 tile (approx. 100 km x 100 km).

Preparation:

  1. Download the demo data which contains Sentinel-2 images for one tile:

    bash gchm/bash/download_demo_data.sh ./
    

    This creates the following directory:

    deploy_example/
    ├── ESAworldcover
    │   └── 2020
    │       └── sentinel2_tiles
    │           └── ESA_WorldCover_10m_2020_v100_32TMT.tif
    ├── image_paths
    │   └── 2020
    │       └── 32TMT.txt
    ├── image_paths_logs
    │   └── 2020
    ├── predictions_provided
    │   ├── 2020
    │   │   ├── S2A_MSIL2A_20200623T103031_N0214_R108_T32TMT_20200623T142851_predictions.tif
    │   │   ├── S2A_MSIL2A_20200623T103031_N0214_R108_T32TMT_20200623T142851_std.tif
    │   │   ├── ...
    │   ├── 2020_merge
    │   │   └── preds_inv_var_mean
    │   │       ├── 32TMT_pred.tif
    │   │       └── 32TMT_std.tif
    │   └── 2020_merge_logs
    │       └── preds_inv_var_mean
    │           └── 32TMT.txt
    ├── sentinel2
    │   └── 2020
    │       ├── S2A_MSIL2A_20200623T103031_N0214_R108_T32TMT_20200623T142851.zip
    │       ├── ...
    └── sentinel2_aws
        └── 2020
    
  2. Download the trained model weights:

    bash gchm/bash/download_trained_models.sh ./trained_models
    

    This creates the following directory:

    trained_models/
    └── GLOBAL_GEDI_2019_2020
        ├── model_0
        │   ├── FT_Lm_SRCB
        │   │   ├── args.json
        │   │   ├── checkpoint.pt
        │   │   ├── train_input_mean.npy
        │   │   ├── train_input_std.npy
        │   │   ├── train_target_mean.npy
        │   │   └── train_target_std.npy
        │   ├── args.json
        │   ├── checkpoint.pt
        │   ├── train_input_mean.npy
        │   ├── train_input_std.npy
        │   ├── train_target_mean.npy
        │   └── train_target_std.npy
        ├── model_1
        │   ├── ...
        ├── model_2
        │   ├── ...
        ├── model_3
        │   ├── ...
        ├── model_4
        │   ├── ...
    

    The checkpoint.pt files contain the model weights. The subdirectories FT_Lm_SRCB contain the models finetuned with a re-weighted loss function.

Deploy example for a single Sentinel-2 image

This demo script processes a single image (from the year 2020) for the tile "32TMT" in Switzerland. Run:

bash gchm/bash/deploy_example.sh

Deploy and merge example for multiple images of a Sentinel-2 tile

This demo script processes 10 images (from the year 2020) for the tile "32TMT" in Switzerland and aggregates the individual per-image maps to a final annual map.

Provide a text file with the image filenames per tile saved as ${TILE_NAME}.txt. The demo data contains the following file:

cat ./deploy_example/image_paths/2020/32TMT.txt 
S2A_MSIL2A_20200623T103031_N0214_R108_T32TMT_20200623T142851.zip
S2A_MSIL2A_20200723T103031_N0214_R108_T32TMT_20200723T142801.zip
S2A_MSIL2A_20200812T103031_N0214_R108_T32TMT_20200812T131334.zip
...

The corresponding images are stored in ./deploy_example/sentinel2/2020/.

  1. Set the paths in gchm/bash/config.sh
  2. Set the tile_name in gchm/bash/run_tile_deploy_merge.sh
  3. Run the script:
    bash gchm/bash/run_tile_deploy_merge.sh
    

Note on ESA World Cover post-processing:

The ESA WorldCover 10 m 2020 v100 reprojected to Sentinel-2 tiles is available on Zenodo.
We apply minimal post-processing and mask out built-up areas, snow,
ice and permanent water bodies, setting their canopy height to ”no data” (value: 255). See the script here.

Note on AWS:

Sentinel-2 images can be downloaded on the fly from AWS S3 by setting GCHM_DOWNLOAD_FROM_AWS="True"
and providing the AWS credentials as described above.
This was tested for 2020 data, but might need some update in the sentinelhub routine to handle newer versions.

Training

Data preparation

  1. Download the train-val h5 datasets from here.
  2. Merge the parts file to a single train.h5 and val.h5 by running this script.
    Before running it, set the variables in_h5_dir_parts and out_h5_dir to your paths. Then run:
    bash gchm/preprocess/run_merge_h5_files_per_split.sh`
    

Running the training script

A slurm training script is provided and submitted as follows.
Before submitting, set the variable CODE_PATH at the top of the script and set the paths in gchm/bash/config.sh. Then run:

sbatch < gchm/bash/run_training.sh

ALS preprocessing for independent comparison

In cases where rastered high-resolution canopy height models are available (e.g. from airborne LIDAR campaigns) for independent evaluation, some preprocessing steps are required to
make the data comparable to GEDI canopy top height estimates corresponding to the canopy top within a 25 meter footprint.

  1. A rastered canopy height model with a 1m GSD should be created (E.g. using gdalwarp).
  2. The 1m canopy height model can then be processed with a circular max pooling operation to approximate "GEDI-like" canopy top heights. This step is provided as a pytorch implementation.

Example:
Download the example CHM at 1m GSD from here. Then run:

python3 gchm/preprocess/ALS_maxpool_GEDI_footprint.py "path/to/input/tif" "path/to/output/tif"

Citation

Please cite our paper if you use this code or any of the provided data.

Lang, N., Jetz, W., Schindler, K., & Wegner, J. D. (2023). A high-resolution canopy height model of the Earth. Nature Ecology & Evolution, 1-12.

@article{lang2023high,
  title={A high-resolution canopy height model of the Earth},
  author={Lang, Nico and Jetz, Walter and Schindler, Konrad and Wegner, Jan Dirk},
  journal={Nature Ecology \& Evolution},
  pages={1--12},
  year={2023},
  publisher={Nature Publishing Group UK London}
}

Owner metadata


GitHub Events

Total
Last Year

Committers metadata

Last synced: 6 days ago

Total Commits: 5
Total Committers: 2
Avg Commits per committer: 2.5
Development Distribution Score (DDS): 0.2

Commits in past year: 5
Committers in past year: 2
Avg Commits per committer in past year: 2.5
Development Distribution Score (DDS) in past year: 0.2

Name Email Commits
nlang n****g@g****h 4
Nico Lang n****l@g****m 1

Committer domains:


Issue and Pull Request metadata

Last synced: 1 day ago

Total issues: 8
Total pull requests: 0
Average time to close issues: about 1 month
Average time to close pull requests: N/A
Total issue authors: 6
Total pull request authors: 0
Average comments per issue: 1.0
Average comments per pull request: 0
Merged pull request: 0
Bot issues: 0
Bot pull requests: 0

Past year issues: 2
Past year pull requests: 0
Past year average time to close issues: 20 days
Past year average time to close pull requests: N/A
Past year issue authors: 2
Past year pull request authors: 0
Past year average comments per issue: 0.0
Past year average comments per pull request: 0
Past year merged pull request: 0
Past year bot issues: 0
Past year bot pull requests: 0

More stats: https://issues.ecosyste.ms/repositories/lookup?url=https://github.com/langnico/global-canopy-height-model

Top Issue Authors

  • Rur1sama (3)
  • andreibrinzea (1)
  • CarlosGomes98 (1)
  • ftong123 (1)
  • ezekielbarnett (1)
  • oguzhannysr (1)

Top Pull Request Authors


Top Issue Labels

Top Pull Request Labels


Dependencies

requirements.txt pypi
  • botocore *
  • ipython *
  • matplotlib *
  • numpy *
  • pathlib *
  • scikit-image *
  • scikit-learn *
  • sentinelhub ==3.9.0
  • tables *
  • tensorboard *
  • tqdm *
  • typing *
  • urllib3 *
  • wandb *
setup.py pypi
environment.yml pypi

Score: 5.786897381366708