Skip to content

🍿POPCORN: High-resolution Population Maps Derived from Sentinel-1 and Sentinel-2 πŸŒπŸ›°οΈ

License

Notifications You must be signed in to change notification settings

prs-eth/Popcorn

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

43 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

POPCORN LOGO

🍿POPCORN: High-resolution Population Maps Derived from Sentinel-1 and Sentinel-2 πŸŒπŸ›°οΈ

Nando MetzgerπŸ¦πŸ“§, Rodrigo Caye Daudt🏦, Devis TuiaπŸ‡, Konrad Schindler🏦

🏦 Photogrammetry and Remote Sensing, ETH Zürich
πŸ‡ Environmental Computation Science and Earth Observation Laboratory, EPFL Sion
πŸ“§ Corresponding Author: nando.metzger@geod.baug.ethz.ch

Website Paper EE_Demo EE_Code EE_Code EE_Code

πŸ’‘ TL;DR

POPCORN is a lightweight population mapping method using free satellite images and minimal data, surpassing existing accuracy and providing interpretable maps for mapping populations in data-scarce regions.

News πŸ“°

  • 14th May 2024: Predictions for Switzerland (che), Rwanda (rwa), and Puerto Rico (pricp2) are now downloadable in .tif format. Download the data here
  • 12th May 2024: We updated the code base with our sparse head implementation. All experiments can now be run with <24GB GPU Memory. See #4.
  • 8th May 2024: We published the training code. View Code on GitHub
  • 20th March 2024: We published the evaluation code and the pretrained models. View Code on GitHub
  • 17th March 2024: Website is live. Visit Website

Abstract πŸ”

Detailed population maps play an important role in diverse fields ranging from humanitarian action to urban planning. Generating such maps in a timely and scalable manner presents a challenge, especially in data-scarce regions. To address it we have developed POPCORN, a population mapping method whose only inputs are free, globally available satellite images from Sentinel-1 and Sentinel-2; and a small number of aggregate population counts over coarse census districts for calibration. Despite the minimal data requirements our approach surpasses the mapping accuracy of existing schemes, including several that rely on building footprints derived from high-resolution imagery. E.g., we were able to produce population maps for Rwanda with 100m GSD based on less than 400 regional census counts. In Kigali, those maps reach an R^2 score of 66% w.r.t. a ground truth reference map, with an average error of only 10 inhabitants/ha. Conveniently, POPCORN retrieves explicit maps of built-up areas and of local building occupancy rates, making the mapping process interpretable and offering additional insights, for instance about the distribution of built-up, but unpopulated areas (e.g., industrial warehouses). Moreover, we find that, once trained, the model can be applied repeatedly to track population changes; and that it can be transferred to geographically similar regions with only a moderate loss in performance (e.g., from Uganda to Rwanda). With our work we aim to democratize access to up-to-date and high-resolution population maps, recognizing that some regions faced with particularly strong population dynamics may lack the resources for costly micro-census campaigns.

The population map of Rwanda is available on Google Earth Engine:

EE DEMO

var popDensity = ee.Image("projects/ee-nandometzger/assets/POPCORNv3");

EE DEMO

Bunia Time Series

Methodology πŸ§ πŸ›°οΈπŸ’»

The core of our method is a neural network model, termed POPCORN. That model has two components: (1) a pre-trained, frozen built-up area extractor; and (2) a building occupancy module that we train through weak supervision with coarse census counts, as illustrated in the Figure below.

The model operates at the full Sentinel-1/-2 resolution, i.e., its output has a nominal spatial resolution of 10m. However, for the final product and evaluation, we recommend aggregating the raw output to a 1ha (100x100m) grid, as done for the evaluation of the paper.

Graphical Abstract

Setup πŸ”§πŸ’Ύ

Environment 🐍

Model training and testing

Instructions on how to install the project or library (Tested on Linux)

Set up the base environment like this:

python -m venv PopMapEnv
source PopMapEnv/bin/activate
pip install -r requirements.txt
pip install torch==2.1.1 torchvision==0.16.1 --index-url https://download.pytorch.org/whl/cu118

Code was tested on Ubuntu 22.04 LTS, 64GB RAM, NVIDIA GeForce RTX 3090 Ti.

Dataset reproducability

If you plan to use the preprocessing tools in this reposiotry, you also need to install GDAL. An easy way to install GDAL without sudo access is as follows:

  • download the gdal-3.4.1 binary, and extract it.
  • install GDAL using these commands (this might take some time):
./autogen.sh
./configure
make

Data πŸŒπŸ—‚οΈ

The code repository contains all the necessary functionalities to reproduce the dataset from the raw data and Google Earth Engine. For the user's convenience, we host necessary preprocessed datasets here. Download and place the data into the following folder structure for Switzerland (che), Rwanda (rwa), and Puerto Rico (pricp2), and make sure to append your data root path in the variables at constants.py.

PopMapData/
β”œβ”€β”€ raw/
β”‚   └── ... (only needed to recompute the dataset processing)
β”œβ”€β”€ processed/
β”‚   β”œβ”€β”€ che
β”‚   β”‚   β”œβ”€β”€census_fine.csv
β”‚   β”‚   β”œβ”€β”€boundaries_fine.tif
β”‚   β”‚   β”œβ”€β”€census_coarse4.csv
β”‚   β”‚   β”œβ”€β”€boundaries_coarse4.tif
β”‚   β”œβ”€β”€ rwa
β”‚   β”‚   β”œβ”€β”€census_coarse.csv
β”‚   β”‚   β”œβ”€β”€boundaries_coarse.tif
β”‚   β”‚   β”œβ”€β”€census_kigali100.csv
β”‚   β”‚   β”œβ”€β”€boundaries_kigali100.tif
β”‚   β”‚   └── ...
β”‚   └── pricp2
β”‚       └── ...
└── merged/
    └── EE/
        β”œβ”€β”€ che
        β”‚   β”œβ”€β”€ S1spring
        β”‚   β”‚   └──rwa_S1spring.tif
        β”‚   β”œβ”€β”€ S1summer
        β”‚   β”‚   └──rwa_S1summer.tif
        β”‚   β”œβ”€β”€ S1autumn
        β”‚   β”‚   └──rwa_S1autumn.tif
        β”‚   β”œβ”€β”€ S1winter
        β”‚   β”‚   └──rwa_S1winter.tif
        β”‚   β”œβ”€β”€ S2Aspring
        β”‚   β”‚   └──rwa_S2Aspring.tif
        β”‚   β”œβ”€β”€ S2Asummer
        β”‚   β”‚   └──rwa_S2Asummer.tif
        β”‚   └── ...
        β”œβ”€β”€ rwa
        β”‚   └── ...
        └── pricp2
            └── ...

Testing πŸ§ͺπŸ—ΊοΈ

Checkpoints πŸ’Ύ

Checkpoints can be downloaded from here.

Inference πŸš€πŸ“Šβš–οΈ

Make sure to add you datapath in the utils/constants.py file. You can now use the run_eval.py script to generate maps and evaluate them subsequently using

python run_eval.py -occmodel -senbuilds -S2 -NIR -S1 -treg <inference dataset name> --fourseasons \
  --resume \
    /path/to/model1/last_model.pth \
    /path/to/model2/last_model.pth \
    ....

The outputs will be written into the folder of the first model. /path/to/model1/last_model.pth in the case above

Training πŸ‹οΈβ€β™‚οΈ

Note: The training script loads large chunks of data per sample. We recommend to use SSDs for data storing to avoid bottlenecks.

Train Switzerland (Estimated training time: ~15h):

python run_train.py -S2 -NIR -S1 -treg che -tregtrain che -occmodel -wd 0.0000005 -senbuilds -pret --biasinit 0.2267

Train Rwanda projected census 2020 (Estimated training time: ~8h):

python run_train.py -S2 -NIR -S1 -treg rwa -tregtrain rwa -occmodel -wd 0.00001 -senbuilds -pret --biasinit 0.9407

Train Puerto Rico (Estimated training time: ~6h):

python run_train.py -S2 -NIR -S1 -treg pricp2 -tregtrain pricp2 -occmodel -wd 0.0000005 -senbuilds -pret --biasinit 0.4119

Uganda:

python run_train.py -S2 -NIR -S1 -treg uga -tregtrain rwa2022 -occmodel -wd 0.00001 -senbuilds -pret --biasinit 0.9407

For the results in the paper, we trained the Bag-of-POPCORN with parameter settings --seed {1600,1601,1602,1603,1604}, default is 1600.

Recompute the dataset πŸ–₯️

To ensure full reproducibility and additional expandability of our workflow. We provide the full data pipeline to recompute the input images:

1. Connect to Google Earth Engine

Make sure you have the gcloud application installed. You need a Google Earth Engine account for this.

Local Machine

If you are on a local machine, you can log in via this command, which will prompt the login page on your browser.

gcloud auth application-default login

Remote Machine (SSH)

If you are on a remote machine, make sure gcloud is installed on the local as well as the remote device. Connect via ssh to you remote machine and run the following command on your remote terminal:

gcloud auth application-default login --no-browser

This will generate another gcloud command like gcloud auth application-default login --remote-bootstrap="....". Copy this command and paste it into your local terminal. Accept that you are bootstrapping glcoud to a trusted machine, and the Earth Engine login window in your browser should be prompted. After successful browser authentification, your local terminal should provide an output https://localhost:8085/.... Copy and paste this line into your remote terminal.

2. Download raw data

python utils/01_download_gee_country.py 28.782241 -2.903950 30.961654 -0.994897 rwa
python utils/01_download_gee_country.py 5.855713 45.759859 10.656738 47.864774 che
python utils/01_download_gee_country.py -67.282031 17.874492 -65.205615 18.522873 pricp2
python utils/01_download_gee_country.py 29.492798 -1.554375 35.095825 4.291636 uga

The resulting files will appear in your google drive.

3. Merging Google Earth Engine outputs

For large regions, GEE will return individual tiles of the scenes. You can merge them together with the utils/03_merge_tiffs.py script. We recommend placing the tiles in the raw/EE/<region>/<modality> folders, where the modality is S1spring or S2Awinter for example. You can then execute the following commands to perform the merging:

python utils/03_merge_tiffs.py <path to data>/PopMapData/raw/EE/rwa <path to data>/PopMapData/merged/EE/rwa 
python utils/03_merge_tiffs.py <path to data>/PopMapData/raw/EE/che <path to data>/PopMapData/merged/EE/che 
python utils/03_merge_tiffs.py <path to data>/PopMapData/raw/EE/pricp2 <path to data>/PopMapData/merged/EE/pricp2 
python utils/03_merge_tiffs.py <path to data>/PopMapData/raw/EE/uga <path to data>/PopMapData/merged/EE/uga 

Note: This process applies a lossless compression, but the outputs can still be quite large. There is a automatic functionality in the dataloader to create virtual files, in case merged files cannot be created here.

4. Census files preprocessing

βš™οΈ Release of the census preprocessing scripts is coming soon.

Citation πŸŽ“

@article{metzger2023high,
  title={High-resolution Population Maps Derived from Sentinel-1 and Sentinel-2},
  author={Metzger, Nando and Daudt, Rodrigo Caye and Tuia, Devis and Schindler, Konrad},
  journal={arXiv preprint arXiv:2311.14006},
  year={2023} 
}

Fun fact

  • "POPCORN" stands for POPulation from COaRrse census Numbers🍿.
  • POPCORN is the successor of POMELO

Releases

No releases published

Packages

No packages published

Languages