Urban flood mapping is challenging due to the complicated structures in cities, such as buildings, sidewalks, road culverts, and utility holes. Two types of satellite products are available for producing flood maps using remote sensing data, including optical and SAR images. Optical images are not always available during a flood event because they are affected by dense cloud cover. SAR sensors, however, can capture images from the flood-affected areas at longer electromagnetic wavelengths making it possible for the SAR signal to penetrate within the cloud cover. Besides, since the SAR sensor is active, it is not dependent on the sunlight. So, SAR images are available in all weather conditions and during day and night. This characteristic makes SAR data suitable for flood mapping.
Generally, three specific features can be extracted from a SAR image, intensity, polarimetry decompositions, and InSAR coherence. Intensity is a measure of the reflective strength of an object, polarimetry decomposition gives the polarimetric discriminators that can be used for classification and image interpretation, and interferometric coherence (correlation) is a measure for the accuracy of the determined radar signal, and its value decreases by temporal changes. While intensity reflects the electromagnetic characteristics of the radar backscattering, it is affected by the speckle noise. Besides, flood mapping based solely on intensity data might result in erroneous flood maps because of the complex structures in urban areas. Vertical structures, like buildings, can enhance the double bounce scattering effect, which can be further intensified when the floodwater covers the bottom of the tree. Coherence data are complementary in flood mapping studies using the SAR dataset (Pulvirenti et al., 2021; Olthof and Svacina, 2020), and the coherency maps show high values in urban areas because of the stability of urban structures during short time intervals. When producing a flood map, a coherency map can complement the intensity data and improve flood mapping accuracy (Zhang et al., 2021). Besides, steady targets such as buildings make InSAR coherence data useful for urban areas with limited vegetated regions. When vegetation cover is limited, any decrease in coherence values in an InSAR time series can be translated into a flood event. Also, speckle noise is reduced when producing coherency images because the noise is averaged when integrating the two SAR images. As mentioned before, because of the dynamic behaviour of the vegetated areas (due to growth), it is not evident that the coherence change is related to the vegetated areas or flood. Sometimes this problem is addressed using SAR images with a short revisit time, less than five days, like COSMO-SkyMed, but such datasets are not accessible quickly, especially for flood hazard management studies in which time plays a vital role (Pierdicca et al. 2018). Another limitation when using SAR data for flood extent mapping in urban areas is the shadowing effect. The shadowing effect in a SAR image happens when the SAR signal does not reach some regions because higher objects create an obstacle between the SAR antenna and the area (Bouvet et al. 2018). The shadowed areas on the image are overlooked when performing flood extent mapping using SAR data.
Flood extent mapping techniques can be categorized into four groups based on the theories applied; 1- Hydrologic/Hydraulic modelling 2- Multi-Criteria Decision Analysis (MCDA) 3- Machine Learning 4- Hybrid methods. Hydrologic models can simulate runoff values during a flood event, and Hydraulic models provide information on flow movement and inundation depth in areas near a river network. Multi-criteria techniques assign a weight to each flood indicator, such as topographic, hydrologic, climatic, and anthropogenic parameters, to produce a final flood risk map. Machine learning approaches, aka Artificial Intelligence techniques, use training data to discriminate between flooded and dry areas based on geospatial input features. Hybrid techniques use a combination of previously mentioned methods to model flood events, such as integrating hydraulic modelling and the Analytical Hierarchical Process (AHP) technique to produce a flood risk map (Nguyen et al. 2021).
Deep Learning, aka deep structured learning, is a machine learning technique based on artificial neural networks with representation learning. Although ML algorithms such as neural networks, random forest, and support vector machines have proven promising methods for flood mapping, DL methods, especially CNNs, have shown higher capability than the previous ML methods to extract features at different scales such as edges and objects (Muñoz et al., 2021). Li et al. (2019a) introduced a CNN to produce a flood map in Houston, USA, during Hurricane Harvey in August 2017 based on TerraSAR-X intensity and coherence data. This study focused on fluvial flooding, and its efficiency for coastal or pluvial flooding was not examined. Some DL-based segmentation models such as Unet, Unet++, and DeepLabV3 have been proposed in the literature for flood mapping, and they have achieved promising results on both optic and SAR images. Wang et al. 2021, proposed a DL model based on Unet for flood water extraction in Poyang Lake in China using Sentinel-1 SAR images. Jaisakthi et al. 2021, proposed a modified Unet algorithm for flood detection using Planet Scope images and reported an overall accuracy of about 70% on validation data. The flood masks in this work were not compared with any ground truth dataset. Konapala et al. 2021, used Unet for flood inundation mapping using Sen1Floods11 data, including Sentinel-1 and Sentinel-2 images from 11 flood events around the globe. After adding elevation data to the input, the flood median F1 score improved from 0.62 to 0.73 compared with using only Sentinel-1 bands. Chen et al. 2022 proposed a Siamese Network based on Unet for building change detection in very high-resolution remote sensing images and reported promising accuracies after comparison with ground truth data. Their method was not tested for flood-induced changes in satellite images.
Convolution Siamese Network (CSN) is one type of DL algorithm that has been applied for change detection (Yang et al. 2021; Wang et al. 2020; Chen et al. 2020). This method highlights changed areas using a bi-temporal remote sensing dataset. CSNs use two parallel CNN in their internal architecture and are used in change detection problems. In CSN, one CNN focuses on the pre-event image and the other works on the co-event image. In this way, CSN is more applicable for change detection problems (here flood mapping) than the usual CNN network. Some recent studies have used Siamese Networks for remote sensing change detection applications. For example, Jian and Li (2021) applied a Siamese Network called S3N. This network used Visual Geometry Group (VGG) as subnetworks and was applied to detect changes in various types of remote sensing data, including panchromatic, MS, SAR, PolSAR and NDVI images. The problem of high computational cost and lack of training data was addressed by applying the transfer learning strategy. They concluded that their proposed architecture is more computationally efficient than state-of-the-art techniques while giving comparable results to the existing methods. Wang et al. (2021) presented a fully CSN trained on Focal Contrastive Loss (FCL) to address the imbalanced data problem by focusing on the samples with fewer train data. Zhang et al. (2022) proposed a Siamese Residual Multi-kernel Pooling module (SRMP) to improve the high-level change information extraction from optical images. A feature difference module was also proposed to extract low-level features and help the model generate more accurate details. In another work, a Siamese Segmentation Network, SiHDNet, was proposed for building change detection. The proposed method was based on deep, high-resolution differential feature interaction. The difference map was created through a special fusion module to obtain sufficient and effective change information. The final binary change map was acquired through the improved spatial pyramid pooling module (Liang et al. 2022). Yang et al. (2021) proposed a new change detection algorithm based on the Siamese Network, MRA-SNet, for building, road, and land cover change detection in optical remote sensing images. The UNet network was used as the backbone architecture, and the bitemporal images were imported separately to the encoder. The ordinary convolution blocks were replaced with Multi-Res blocks to extract spatial and spectral features of different scales in remote sensing images. These studies however were all based on optical image data for change detection, and they did not address the challenges associated with the SAR change detection problem. Recently, Siamese Networks have been used for flood mapping studies. Zhang et al. (2022) proposed a domain adaptation-based multi-source change detection method for heterogeneous remote sensing images. The Landsat-8 image was used as a pre-event, and the Sentinel-1A image was used as the co-event for flood mapping of the 2017 California event. The area studied for flood mapping in this work covered agricultural lands, not dense urban areas.
In this study, flood extent mapping is considered a change detection problem, thus CSN is applied to discern between flooded areas and non-flood regions. The contribution of this research is the use of the SAR data and a deep learning-based change detection algorithm, Convolutional Siamese Network (CSN), for urban flood mapping. Because of the SAR limitations, including geometric distortions such as layover, shadowing effects, and speckle noise, the use of SAR data for flood mapping is already a challenging task. In addition, deep learning for flood mapping can be a challenge because of the high computational burden that these algorithms add to the process. Hence, the use of SAR and deep learning algorithm for flood mapping is examined in this study. Producing flood maps in urban areas is more challenging than in rural or open areas because of the complex structures in the cities. To the best of our knowledge, this is the first study using SAR data and CSN for urban flood mapping.
The structure of this paper is as follows, Section 2 presents the study area and dataset applied. Section 3 describes the methodology and Section 4 discusses the results. Section 5 presents a discussion of the obtained results, followed by the conclusion in Section 6.