Skip to content

Latest commit

 

History

History
69 lines (49 loc) · 5.6 KB

File metadata and controls

69 lines (49 loc) · 5.6 KB

Paper

Last update: May 31, 2023

Abstract

Rapid response to natural hazards, such as floods, is essential to mitigate loss of life and the reduction of suffering. For emergency response teams, access to timely and accurate data is essential. Satellite imagery offers a rich source of information which can be analysed to help determine regions affected by a disaster. Much remote sensing flood analysis is semi-automated, with time consuming manual components requiring hours to complete. In this study, we present a fully automated approach to the rapid flood mapping currently carried out by many non-governmental, national and international organisations. We take a Convolutional Neural Network (CNN) based approach which isolates the flooded pixels in freely available Copernicus Sentinel-1 Synthetic Aperture Radar (SAR) imagery, requiring no optical bands and minimal pre-processing. We test a variety of CNN architectures and train our models on flood masks generated using a combination of classical semi-automated techniques and extensive manual cleaning and visual inspection. Our methodology reduces the time required to develop a flood map by 80%, while achieving strong performance over a wide range of locations and environmental conditions. Given the open-source data and the minimal image cleaning required, this methodology can also be integrated into end-to-end pipelines for more timely and continuous flood monitoring.

Tables and Figures

Hyper-parameter tuning experimental setups for different architectures. Here we vary thebatch size, number of times the training dataset is passed to the model during training, the use of aweighted loss function, the filter depth of the convolutional layers (‘deep’ is four times the depth of‘shallow’) and the use of mixed precision during training. For the XNet and U-Net models this number refers to the number of times the entire training dataset was passed to themodel during training before Early Stopping was implemented. For the U-Net+ResNet case, the first number is the numberof times the dataset was passed when training just the head nodes of the architecture, and the latter the number of times itwas passed when training the entire network. In the latter case, at each epoch the validation loss was manualy comparedafter training and the best model selected.

Name Batch size Dataset passes Weighted Loss Filter depth Mixed
XNet shallow 3 1 [16,32,64,128]
XNet shallow weighted 3 1 x [16,32,64,128]
XNet shallow weighted b8 8 1 x [16,32,64,128]
XNet shallow weighted b8 d5 8 5 x [16,32,64,128]
XNet shallow weighted 3 1 x [56,128,256,512]
UNet shallow 3 1 [16,32,64,128]
UNet shallow weighted 3 1 x [16,32,64,128]
UNet shallow weighted d5 3 5 x [16,32,64,128]
UNet shallow weighted b8 8 1 x [16,32,64,128]
UNet shallow weighted b8 d5 8 5 x [16,32,64,128]
UNet shallow weighted 3 1 x [56,128,256,512]
UNet + resNet b4 d5/10 4 5/10 x [64,128,256,512,1204] x
UNet + resNet b8 d5/10 8 5/10 x [64,128,256,512,1204] x
UNet + resNet b8 d5/10 8 5/10 x [64,128,256,512,1204]
UNet + resNet b8 d10/20 8 10/20 x [64,128,256,512,1204] x
UNet + resNet b32 d10/200 32 10/20 x [64,128,256,512,1204] x

Precision-recall curved for different experiments from top to bottom. XNet trials; U-Net trials; and U-Net+ResNet trials. The curves show the precision and recall values at different probability thresholdsranging from 0.01 to 0.99.

Overall quantitative comparison of the best XNet, U-Net and U-Net+ResNet models after hyper-parameter tuning compared against the automatic histogram based method baseline result.

Model Accuracy Precision Recall Critical Sucess Index F1
Baseline 91% 62% 84% 0.55 0.71
XNet 97% 91% 91% 0.81 0.91
UNet 97% 91% 92% 0.83 0.91
U-Net + ResNet 97% 91% 92% 0.77 0.92

Example outputs of the best performing U-Net+ResNet, after probability thresholding, can be seen beloe. In particular, we see the neural network's ability to detect the flood area with minimal cleaning in comparison to the ground-truth data. The baseline (third column) was generated using the automatic threshold-based method and would require significantly more noise reduction in post-processing.

From left to right: tiles of different analysis are shown with the corresponding ground truth,baseline and neural network prediction.

Examples of well detected tiles from both the automatic histogram-based method and this neural network, particularly highlighting severe flooded regions: