Skip to content

Latest commit

 

History

History
66 lines (46 loc) · 2.28 KB

README.md

File metadata and controls

66 lines (46 loc) · 2.28 KB

Decloud

Decloud enables the training and inference of various neural networks to remove clouds in optical images.

Representative illustrations:

Examples of de-clouded Sentinel-2 images using the single date SAR/Optical U-Net model.

Cite

@inproceedings{cresson2022comparison,
  title={Comparison of convolutional neural networks for cloudy optical images reconstruction from single or multitemporal joint SAR and optical images},
  author={Cresson, R{\'e}mi and Nar{\c{c}}on, N and Gaetano, Raffaele and Dupuis, Aurore and Tanguy, Yannick and May, St{\'e}phane and Commandr{\'e}, Benjamin},
  booktitle={XXIV ISPRS Congress (2022 edition)},
  volume={43},
  pages={1317--1326},
  year={2022}
}

https://doi.org/10.48550/arXiv.2204.00424

Quickstart: Run a pre-trained model

Some pre-trained models are available at this url.

The easiest way to run a model is to run the timeseries processor such as:

python production/meraner_timeseries_processor.py
        --s2_dir  S2_PREPARE/T31TCJ 
        --s1_dir  S1_PREPARE/T31TCJ
        --model   merunet_occitanie_pretrained/
        --dem     DEM_PREPARE/T31TCJ.tif
        --out_dir meraner_timeseries/
# Optional arguments:
        --write_intermediate --overwrite
        --start 2018-01-01 --end 2018-12-31 
        --ulx 306000 --uly 4895000 --lrx 320000 --lry 4888000

You can find more info on available models and how to use these models here

Advanced usage: Train you own models

  1. Prepare the data: convert Sentinel-1 and Sentinel-2 images in the right format (see the documentation).
  2. Create some Acquisition Layouts (.json files) describing how the images are acquired, ROIs for training and validation sites, and generate some TFRecord files containing the samples.
  3. Train the network of your choice. The network keys for input/output must match the keys of the previously generated TFRecord files.
  4. Perform the inference on real world images.

More info here.

Contact

You can contact remi cresson (Remi Cresson at INRAE )