tri-band assessment of multi-spectral satellite data for flood...

10
Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detection Pallavi Jain, Bianca Schoen-Phelan, Robert Ross Technological University Dublin, Dublin, Ireland {pallavi.jain, bianca.schoenphelan, robert.ross}@tudublin.ie Abstract. Multi-spectral satellite data provides vast resources for im- portant tasks such as flood detection, but training and fine tuning mod- els to perform optimally across multi-spectral data remains a significant research challenge. In light of this problem, we present a systematic ex- amination of the role of tri-band deep convolutional neural networks in flood prediction. Using Sentinel-2 data we explore the suitability of different deep convolutional architectures in a flood detection task; in particular we examine the utility of VGG16, ResNet18, ResNet50 and EfficientNet. Importantly our analysis considers the questions of different band combinations and the issue of pre-trained versus non-pre-trained model application. Our experiment shows that a 0.96 F1 score is achiev- able for our task through appropriate combinations of spectral bands and convolutional neural networks. For flood detection, three-band com- binations of RB8aB11 and RB11B outperformed 33 other combinations when trained with pre-trained ResNet18 and other models. Our anal- ysis further demonstrates a strong performance by pre-trained models despite the fact that these pre-trained models were originally trained on different spectral bands. Keywords: Remote Sensing · Deep Convolutional Neural Network · Multi-spectral · Flood Detection · Sentinel-2. 1 Introduction Floods are natural hazards that occur throughout the year in many different parts of the world, and can occur for reasons including heavy rainfall, melting snow or tsunamis. Floods damages properties and agricultural areas, and are also highly hazardous to human life [10]. These factors together highlight the need for timely and accurate detection in order to make flood management operation more effective. To this end satellite data processing has become a vital diagnostic tool. Satellite data provides much more than simple RGB (Red-Green-Blue) im- age data. Instead satellites are multi-spectral instruments (MSI) that generate multi-spectral and sometimes hyper-spectral data over very different wavebands to RGB. The importance of this is that different spectral bands are reflected or absorbed differently depending on geo-physical properties, e.g., short-wave Copyright © 2020 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).

Upload: others

Post on 27-Jul-2021

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

Tri-Band Assessment of Multi-Spectral SatelliteData for Flood Detection

Pallavi Jain, Bianca Schoen-Phelan, Robert Ross

Technological University Dublin, Dublin, Ireland{pallavi.jain, bianca.schoenphelan, robert.ross}@tudublin.ie

Abstract. Multi-spectral satellite data provides vast resources for im-portant tasks such as flood detection, but training and fine tuning mod-els to perform optimally across multi-spectral data remains a significantresearch challenge. In light of this problem, we present a systematic ex-amination of the role of tri-band deep convolutional neural networksin flood prediction. Using Sentinel-2 data we explore the suitability ofdifferent deep convolutional architectures in a flood detection task; inparticular we examine the utility of VGG16, ResNet18, ResNet50 andEfficientNet. Importantly our analysis considers the questions of differentband combinations and the issue of pre-trained versus non-pre-trainedmodel application. Our experiment shows that a 0.96 F1 score is achiev-able for our task through appropriate combinations of spectral bandsand convolutional neural networks. For flood detection, three-band com-binations of RB8aB11 and RB11B outperformed 33 other combinationswhen trained with pre-trained ResNet18 and other models. Our anal-ysis further demonstrates a strong performance by pre-trained modelsdespite the fact that these pre-trained models were originally trained ondifferent spectral bands.

Keywords: Remote Sensing · Deep Convolutional Neural Network ·Multi-spectral · Flood Detection · Sentinel-2.

1 Introduction

Floods are natural hazards that occur throughout the year in many differentparts of the world, and can occur for reasons including heavy rainfall, meltingsnow or tsunamis. Floods damages properties and agricultural areas, and are alsohighly hazardous to human life [10]. These factors together highlight the needfor timely and accurate detection in order to make flood management operationmore effective. To this end satellite data processing has become a vital diagnostictool.

Satellite data provides much more than simple RGB (Red-Green-Blue) im-age data. Instead satellites are multi-spectral instruments (MSI) that generatemulti-spectral and sometimes hyper-spectral data over very different wavebandsto RGB. The importance of this is that different spectral bands are reflectedor absorbed differently depending on geo-physical properties, e.g., short-wave

Copyright © 2020 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).

Page 2: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

2 P. Jain et al.

infrared (SWIR) bands can discriminate between wet soil and dry soil whereasnear infrared (NIR) wavelength are absorbed by water and reflects by vegetation.This property of MSI provides volumes of information in terms of reflectance atspecific wavelengths across geographic areas.

Observing and mapping floods using satellite image data is an active areaof research with many promising and effective techniques being developed [18,15]. The main challenges using satellite data present in urban areas and in areasof high vegetation coverage. Challenges are due to water having variations inoptical properties across different geographical features such as rivers, lakes,oceans, ponds and floods due to their different compositions, depths and localinteractions. Local composition issues make shallow waters, for example rivers,difficult to detect.

Approaches to the multi-spectral imaging processing have evolved in recentyears. Traditional approaches to multi-spectral data processing rely on hand-crafted features of spectral reluctances. These indexing techniques are known tobe sub-optimal for image processing. Recent trends in Deep Neural Networksprovide an opportunity to learn optimal predictive functions that fully leveragethe potential of available MSI satellite data. While deep networks and CNNs inparticular are highly successful in image classification, they generally requireslarge amount of labelled data. For this reason pre-trained models are of greatuse, and have shown great performances in several domains.

One challenge with pre-trained models is that they have generally beentrained only on 3 channel image data corresponding to Red, Green and Blueinformation. The specific types of features that they have been trained to iden-tify are dependent on combinations of these spectral bands. It is not obviouswhether the features that have been learned are easily applicable to other com-binations of spectral bands such as what is required for flood detection.

Given the innate advantage of CNNs in a task such as flood detection, inthis paper we examine the effectiveness of pre-trained vs non-pre-trained CNNmodels on multi-spectral satellite image data processing. We do this specificallyfor the task of flood detection, using Sentinel-2 data. Three concerns are par-ticularly interesting here: how well do different pre-trained models trained onRGB compare to each other when used for to our multi-spectral flood detectiontask; is there a notable improvements over models trained from scratch on multi-spectral input data; and is there a specific three-band combination of spectralbands that outperforms others (pre-trained models have been trained on threebands which means that three bands need to be input when using such models).Before proceeding to introduce the specifics of our investigation, we first reviewsome key related work.

2 Related Work

It is long known that different wavelengths vary considerably in their reflectancesfor different geo-physical properties such as presence of water bodies or vegeta-tion, and indeed the presence of specific chemicals. Given these reflectance prop-

Page 3: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

Tri-Band Assessment of Multi-Spectral Satellite Data 3

erties, many thresholding techniques [2, 12] and water indexing techniques [13,19, 6] have been proposed in the past to identify pixels corresponding to waterin multi-spectral images.

Since these handcrafted indexing techniques have been developed to detectwater bodies in general, they are not useful for distinguishing flood water frompermanent water bodies. One of the reasons for their inefficiency is that floodwater is usually shallow and possesses a similar reflectance value as built-up areasor cloud shadows, and that their images contain mixed pixel values due to sedi-ments and other objects [3]. Ideally, in order to detect a flood these techniquesrequire pre-flood and post-flood images, which may not be always available.Since water indices are based on the bands capabilities of highlighting certaingeological properties, it is reasonable to analyse different combinations of bands.Additionally, this provides a wider and more robust detection criteria, using au-tomatic detection without any handcrafted features or static thresholding, asthese are highly sensitive to several factors such as region, and similar spectralsignature.

Rather than having to rely on handcrafted features, in recent years it hasbeen shown that deep learning based CNNs provide great performance in sev-eral remote sensing tasks, for example scene classification, building mappingor detection of passable roads [8, 1]. There have been many variants on CNNarchitectures over the last few years with models such as VGG [16], ResNet[7], EfficientNet [17] on the forefront. Unfortunately, there is no universal CNNarchitecture that could be applied to every task, and a model that is state-of-the-art in one domain, is not necessarily guaranteed to be effective in anotherdomain. Since CNNs have the ability to learn spatial as well as spectral fea-tures of the images [20], this is valuable in multi-spectral satellite imaging wherespatial variance is also important. For example, EuroSat data contains a largeamount of high resolution images for land cover classification. ResNet50 is themost successful model on this data, classifying with 98.57% accuracy [8].

Training a residual network requires a large amount of data, which may notbe available. Transfer learning, which means pre-training the network on hugeamount of available datasets such as ImageNet on the ILSVRC challenge [5], andthen adjusting the model by inputting domain specific data, may be the solutionin those cases and have demonstrated very good performance in the past [9, 11].The drawback of transfer learning is that all available pre-trained models usethree channels, i.e. RGB. As multi-spectral satellite data typically offers around10-13 bands, specific three band combinations need to be found that are suitablefor flood detection from satellite images, which is where this work seeks to make acontribution. It is currently unclear whether wavelength dependent deep learningmodels generalise well to other wavelength combinations.

3 Study Design

Below we set out the detail of our study including the selection of data, the selec-tion of candidate CNN models, the training processing, and evaluation methods.

Page 4: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

4 P. Jain et al.

3.1 Data

SENTINEL-2 provides high-resolution, multi-spectral images, and monitors land,vegetation, soil and water cover, as well as observation of inland waterways andcoastal areas. It has multi-spectral instrument (MSI) samples with 13 spectralbands. Among the 13 spectral bands, four bands are at 10 metres, six bands at20 metres and three bands at 60 metres spatial resolution.

We leverage the annotated dataset provided by the MediaEval1 2019 compe-tition [4]. The data consists of 335 image sets with 267 identified as developmentsets and 68 as test sets. Each set consists of between 1 to 24 day time seriesimages of before and after flood events; this provides a total of 2,770 images.We filtered images which did not had full coverage or had full cloud coverage,which made our dataset total with 2180 images. The data has 12 bands insteadof the 13 bands available on SENTINEL-2, which comes in three different setsof resolutions: 10 metres, 20 metres and 60 metres. Each 10 metre resolutionimages 512 X 512 pixels in size, 20 metre resolution images are of 256 X 256pixels, and 60 metre images are of 128 X 128 pixels in size. Among 12 bands, weexcluded 60m band images, which left us with 10 bands in total.

The development dataset is split into three parts: training, validation, andtest, in the ratio of 80:10:10. The split was done based on location that is 267set of images, to remove any learning bias in our test data. Although the dataprovides us with time-series data but we haven’t utilise the temporal informationin our current work. Additionally, we performed image augmentation by shifting,rotating, and flipping the images with batch sizes of 16 and 8 in order to increaseour training dataset and remove any biases.

3.2 Image Pre-Processing and Processing

Sentinel-2 bands Wavelength (nm) Resolution(m)B – Blue 492 10G – Green 559 10R – Red 664 10B5 – Vegetation red edge 704 20B6 – Vegetation red edge 740 20B7 – Vegetation red edge 782 20B8 – NIR 832 10B8a – Narrow NIR 864 20B11 – SWIR 1613 20B12 – SWIR 2202 20

Table 1: Band Notations Used and their Wavelengthand Spatial Resolution

As the reflectance value rangevaries considerably for differentbands, we normalised image’s eachband’s pixel or reflectance value toa standard range from 0-255. Fur-thermore, we up-scaled the 20mresolution band images to 10m res-olution images due to the factthat images were provided withtwo different spatial resolutions.After normalising and up-scalingwe stacked three different bands to-gether to form various three chan-nel combinations. The process ofselecting the three band combina-tions was based on the selection of bands rather than sequence they are stacked.

1 http://www.multimediaeval.org/mediaeval2019/

Page 5: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

Tri-Band Assessment of Multi-Spectral Satellite Data 5

This made 120 combinations from 10 base bands, out of which we showed only33 combinations in this work, whose overall F1 score was greater than thresholdvalue of 75.

Available spectral bands, their names, and shorthand notations for each bandare provided in Table 1; here we used R, G, B as is, while other bands no-tations were used as per their wavelength positions. Given this notation, andthe 10 bands, we analysed 33 band combinations as follows: RB8aB11, RB11B,B7B11B, B7GB11, B8B11R, B6GB, B8aGB, RB8aB, B7GB, B7GB8, RGB11,B7B8aB11, B7B8aB, B8GB, RB8B, RB8aB7, RGB8, B7B8B, B11GB, B7B8B11,RB11B7, RB7B, B8aB11B, B8aGB11, B5GB, B7GB8a, RGB7, RGB8a, RB12B,B8GB11, RB8B7, RGB and B8B11B.

3.3 CNN based Training Models

The basic building block of a CNN is a multi-tier network which includes convo-lution layer, pooling layer, and fully connected layer, which extracts the featuressuch as lines and edges. Over time CNNs have evolved and are now able to learnmore features by increasing the depth and width of the network, which we calla deep CNN. There are many variants of CNN architectures available. We ex-amine three important variants: VGG [16], Deep Residual Network ResNet [7],and EfficientNet [17]. These networks are popular due to their performance inimage classification tasks. All considered networks are very deep and released inseveral depth variants. For example, VGG has an architecture with 16 and anarchitecture with 19 layers.

Fig. 1: Experiment Design

This study uses ImageNet pre-trained variants of VGG16, ResNet18, ResNet50and EfficientNetB0. Global average pooling was used at the output of each pre-trained model, which then fed to a fully connected layer of 512 units with a ReLUactivation function. In order to avoid over-fitting during training, a dropout of0.5 was used. The final output layer used a sigmoid function for binary classifi-cation.

Pre-trained models are compared with their respective versions trained fromscratch on our available dataset, described above. A VGG16, ResNet18 and

Page 6: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

6 P. Jain et al.

ResNet50 models was trained. For both pre-trained and ’from scratch’ models,rectified linear unit (ReLU) has been used as activation function and the Adamoptimiser was applied to guide the training process. Flood detection is a bi-nary classification problem, i.e. flood present and flood not present. Therefore,a binary cross entropy loss function was used in order to calculate the loss. Fortraining of the model an initial learning rate of 5e-6 was used. All pre-trainedmodels were trained for 100 epochs with a batch size of 16. For VGG16, ResNet18and ResNet50 models from scratch, a batch size of 8 was used and training oc-curred for a variable number of epochs, as per their best performance. For thetraining phase an NVIDIA Tesla K40m GPU was used.

3.4 Evaluation Method

For evaluation of the binary classification i.e. flooded and non-flooded region, F1score and Kappa coefficient are used. The F1 score is a harmonic mean of theprecision and recall performance metrics. Cohen’s Kappa meanwhile comparesobserved accuracy with expected accuracy. Kappa provides the fair comparisonwhen the classes are imbalanced as in the case for our data.

4 Results

Results focus on the top 10 waveband combinations, an analysis of the top 2performing results across the pre-trained CNN variants, and finally, an analysisof pre-trained models versus those trained from scratch.

4.1 Three-Band Combinations

Fig. 2: Comparative Results of 33 Combinations from VGG16, ResNet18, ResNet50, and Efficient-NetB0

Page 7: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

Tri-Band Assessment of Multi-Spectral Satellite Data 7

In order to have the best performing combinations for detecting flood events,33 different waveband combinations were run. Although each pre-trained modelhas its own best combination, some combinations performed better overall on allthe models. Figure 2 shows the boxplot result for each combination’s F1 scorespread across the four pre-trained models. Along with boxplot, mean lines areincluded to clearly show the average for each combination. Overall competitiveresults are produced by RB11B, RB8aB11, and B7B11B. While RGB, RB12B,and B8B11B under-perform across all models. This is consistent with previousresearch demonstrating that NIR and SWIR bands both are good at highlightingwater pixels [13, 14, 19]. The performance of RB8aB11 shows that the NNIR bandis also capable of providing good results when combined with Red and SWIRbands. Furthermore, with respect to the performance of RB11B, the Blue bandis useful for mapping depths and shapes of underwater terrain, and distinguishessoil from vegetation. This highlights water pixels in vegetated and urban areasspecifically when combined with the other successful bands.

Fig. 3: Comparison of VGG16, ResNet18,ResNet50, and EfficientNetB0 model’s F1 ScoreAcross All the Combinations

Combination Model F1 Kappa TN TP

RB8aB11 ResNet18 96 91.3 95 97RB11B ResNet18 96 91.2 97 94B7B11B ResNet18 95.4 90.5 95 95RB8aB ResNet50 95.4 90 96 94B11GB ResNet50 95 89 95 94B7GB11 ResNet18 95 89 96 93B8aGB ResNet50 94.5 88.4 95 94RB8B11 ResNet18 94.4 88.2 92 98B7B8B11 VGG16 94.4 88 97 90B7B8aB11 EfficientNetB0 94.3 88 94 95

Table 2: Top-10 Best Performing Combina-tions in Terms of F1 and Kappa

4.2 Pre-Trained Model Performance

Four pre-trained models are run across 33 different three-band combinations:VGG16, ResNet18, ResNet50 and EfficientNetB0. According to figure 3, VGG16and EfficientNetB0 performed relatively poorly in comparison to ResNet18 andResNet50. ResNet50 achieved slightly higher median and average compared toResNet18 but ResNet18 has the more compact spread compared to ResNet50. In-creasing the layers improves the performance. However, ResNet18 and ResNet50perform among the top 3 in terms of accuracy, as illustrated in table 2. Thisindicates that the water identification process may not necessarily require muchdeeper models than 18, which has the added benefit of being detected at earlierstages as the model finishes faster.

4.3 Pre-Trained vs Models from Scratch

Page 8: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

8 P. Jain et al.

Model F1 Kappa EpochResNet18 89 76.5 500ResNet50 93 84 500VGG16 89 76 300VGG16 86.4 71.6 400

Table 3: Results for Models Trainedfrom Scratch on RB8aB11

Given that RB8aB11 showed overall bestresults, VGG16, ResNet18 and ResNet50models were trained from scratch for thisspecific target combination. The number ofepochs varied depending on performance ofthe model. Result are illustrated in table3. Among the three models trained fromscratch, ResNet50 showed the overall best re-sults but it could not compete with pre-trained models. Moreover, ResNet50’saccuracy for training and validation flat-lined after 500 epochs.

As seen earlier in table 2, the top-3 accuracy was achieved by the ResNet18pre-trained model, while ResNet50 shows the best result among the ’from scratch’models. It could be argued that pre-trained models require less depth, whiletraining from scratch requires deeper models in order to achieve the competitiveresults. This also highlights that it could be possible to improve the accuracyby training networks deeper than Resnet50 from scratch in order to competewith pre-trained model performance. This affects computational cost. Amongthe four models VGG16 takes the longest time to train while ResNet18 andEfficientNetB0 were the fastest.

5 Discussion

SWIR is known for its ability to penetrate thin clouds, smoke and haze betterthan visible bands. Considering the low reflectance property of water in SWIRbands, it is perhaps not a surprise that SWIR1 is the most common band seenin the Top-10 accuracy results in table 2. Consequently, SWIR1 (B11) is a goodseparator of water when bands with different or similar reflectance propertiesare combined. It is also noticeable, that not all combinations with SWIR1 (B11)perform well, as B8B11B achieved the poorest result. This could be due to thefact that none of the three bands distinguish vegetation. Instead, they providepredominantly water and cloud identifiers. This also indicates the need for bal-ance among bands in order to highlight a particular feature. Therefore, it iscrucial for further analysis to choose the right combination of bands.

6 Conclusion

This study illustrates that some particular spectral bands combinations excelwith almost all deep CNN models in the flood classification task. The best threecombinations observed were RB8aB11, RB11B and B7B11B. We suggest thatthis is due to their spectral sensitivity to distinguish water and soil/vegetationclearly. Significantly, our analysis also showed that pre-trained models easilyoutperform models trained from scratch. Amongst the examined architecturesResNet18 performed best in the case of pre-trained data, while ResNet50 per-formed best in the case of training models from scratch. On hand hand thisis unsurprising since pre-trained models trained on large volumes of data are

Page 9: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

Tri-Band Assessment of Multi-Spectral Satellite Data 9

(a) RGB (b) RB8aB11 (c) RB11B (d) B7B11B

(e) RGB (f) RB8aB11 (g) RB11B (h) B7B11B

Fig. 4: Pre and Post Flood Images for RGB and Top-3 Band Combination , where (a),(b),(c),(d)are Pre-Flood images and (e),(f),(g),(h) are Post-Flood images

known to give benefit to domains with low volumes of data. On the other handit is somewhat surprising that the differences in wavebands and not only trainingtask does not provide a greater limitation when the pre-trained models are ap-plied to new domains. We suggest that this will be due to the relative dominanceof spatial features rather than specific spectral combinations in the pre-trainednetworks.

One limitation of this study is that only three-band combinations were con-sidered. This is due to our goal of examining pre-trained networks that to thispoint are by definition trained using three channels and consequently expectthree channels as input data. Examining larger combinations of spectral bandsis naturally important. A natural extension of this work is to investigate theuse of actual MSI data in the construction of suitable pre-trained networks thatcan be applied to multi-spectral tasks. While we leave some of these issues forfuture work, we see the current work as a useful contribution in confirming theoverall applicability of pre-trained networks in satellite based imaging tasks suchas flood detection.

References

1. Ahmad, K., Pogorelov, K., Riegler, M., Ostroukhova, O., Halvorsen, P., Conci, N.,Dahyot, R.: Automatic detection of passable roads after floods in remote sensed andsocial media data. Signal Processing: Image Communication 74, 110–118 (2019)

2. Ban, H.J., Kwon, Y.J., Shin, H., Ryu, H.S., Hong, S.: Flood monitoring usingsatellite-based rgb composite imagery and refractive index retrieval in visible andnear-infrared bands. Remote Sensing 9(4), 313 (2017)

3. Bangira, T., Alfieri, S.M., Menenti, M., van Niekerk, A.: Comparing threshold-ing with machine learning classifiers for mapping complex water. Remote Sensing11(11), 1351 (2019)

Page 10: Tri-Band Assessment of Multi-Spectral Satellite Data for Flood Detectionceur-ws.org/Vol-2766/paper8.pdf · 2020. 12. 9. · Table 1: Band Notations Used and their Wavelength and Spatial

10 P. Jain et al.

4. Benjamin Bischke, Patrick Helber, Erkan Basar, Simon Brugman, Zhengyu Zhaoand Konstantin Pogorelov: The Multimedia Satellite Task at MediaEval 2019:Flood Severity Estimation. In: Proc. of the MediaEval 2019 Workshop. Sophia-Antipolis, France (2019)

5. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer visionand pattern recognition. pp. 248–255. Ieee (2009)

6. Feyisa, G.L., Meilby, H., Fensholt, R., Proud, S.R.: Automated water extractionindex: A new technique for surface water mapping using landsat imagery. RemoteSensing of Environment 140, 23–35 (2014)

7. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In:Proceedings of the IEEE conference on computer vision and pattern recognition.pp. 770–778 (2016)

8. Helber, P., Bischke, B., Dengel, A., Borth, D.: Eurosat: A novel dataset and deeplearning benchmark for land use and land cover classification. IEEE Journal ofSelected Topics in Applied Earth Observations and Remote Sensing 12(7), 2217–2226 (2019)

9. Hu, F., Xia, G.S., Hu, J., Zhang, L.: Transferring deep convolutional neural net-works for the scene classification of high-resolution remote sensing imagery. RemoteSensing 7(11), 14680–14707 (2015)

10. Jonkman, S.N., Kelman, I.: An analysis of the causes and circumstances of flooddisaster deaths. Disasters 29(1), 75–97 (2005)

11. Liu, X., Chi, M., Zhang, Y., Qin, Y.: Classifying high resolution remote sensingimages by fine-tuned vgg deep networks. In: IGARSS 2018-2018 IEEE InternationalGeoscience and Remote Sensing Symposium. pp. 7137–7140. IEEE (2018)

12. Matgen, P., Hostache, R., Schumann, G., Pfister, L., Hoffmann, L., Savenije, H.:Towards an automated sar-based flood monitoring system: Lessons learned fromtwo case studies. Physics and Chemistry of the Earth, Parts A/B/C 36(7-8), 241–252 (2011)

13. McFeeters, S.K.: The use of the normalized difference water index (ndwi) in thedelineation of open water features. International journal of remote sensing 17(7),1425–1432 (1996)

14. Mishra, K., Prasad, P.: Automatic extraction of water bodies from landsat imageryusing perceptron model. Journal of Computational Environmental Sciences 2015(2015)

15. Notti, D., Giordan, D., Calo, F., Pepe, A., Zucca, F., Galve, J.P.: Potential andlimitations of open satellite data for flood mapping. Remote sensing 10(11), 1673(2018)

16. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scaleimage recognition. arXiv preprint arXiv:1409.1556 (2014)

17. Tan, M., Le, Q.: Efficientnet: Rethinking model scaling for convolutional neuralnetworks. In: International Conference on Machine Learning. pp. 6105–6114 (2019)

18. Wieland, M., Martinis, S.: A modular processing chain for automated flood moni-toring from multi-spectral satellite data. Remote Sensing 11(19), 2330 (2019)

19. Xu, H.: Modification of normalised difference water index (ndwi) to enhance openwater features in remotely sensed imagery. International journal of remote sensing27(14), 3025–3033 (2006)

20. Yue, J., Zhao, W., Mao, S., Liu, H.: Spectral–spatial classification of hyperspectralimages using deep convolutional neural networks. Remote Sensing Letters 6(6),468–477 (2015)