iBet uBet web content aggregator. Adding the entire web to your favor.
iBet uBet web content aggregator. Adding the entire web to your favor.



Link to original content: https://doi.org/10.3390/s22218245
FAPNET: Feature Fusion with Adaptive Patch for Flood-Water Detection and Monitoring
Next Article in Journal
A 24 GHz CMOS Direct-Conversion RF Receiver with I/Q Mismatch Calibration for Radar Sensor Applications
Next Article in Special Issue
The Detection of Foreign Items in Laundry Industry by Dual-Energy X-ray Transmission—Advantages and Limits
Previous Article in Journal
FDMLNet: A Frequency-Division and Multiscale Learning Network for Enhancing Low-Light Image
Previous Article in Special Issue
Determination of the Strain Influence on the InAs/InAsSb Type-II Superlattice Effective Masses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FAPNET: Feature Fusion with Adaptive Patch for Flood-Water Detection and Monitoring †

1
Multimedia Research Centre, University of Alberta, Edmonton, AB T6G 2E8, Canada
2
3vGeomatics Inc., Vancouver, BC V5Y 0M6, Canada
*
Author to whom correspondence should be addressed.
Islam, M.S.; Sun, X.; Wang, Z.; Ghuman, P.; Cheng, I. MNET: Semantic Segmentation for Satellite Images Based on Multi-Channel Decomposition. Presented at the 9th International Symposium on Sensor Science, Warsaw, Poland, 20–22 June 2022.
Sensors 2022, 22(21), 8245; https://doi.org/10.3390/s22218245
Submission received: 17 September 2022 / Revised: 15 October 2022 / Accepted: 19 October 2022 / Published: 27 October 2022
(This article belongs to the Special Issue I3S 2022 Selected Papers)

Abstract

:
In satellite remote sensing applications, waterbody segmentation plays an essential role in mapping and monitoring the dynamics of surface water. Satellite image segmentation—examining a relevant sensor data spectrum and identifying the regions of interests to obtain improved performance—is a fundamental step in satellite data analytics. Satellite image segmentation is challenging for a number of reasons, which include cloud interference, inadequate label data, low lighting and the presence of terrain. In recent years, Convolutional Neural Networks (CNNs), combined with (satellite captured) multispectral image segmentation techniques, have led to promising advances in related research. However, ensuring sufficient image resolution, maintaining class balance to achieve prediction quality and reducing the computational overhead of the deep neural architecture are still open to research due to the sophisticated CNN hierarchical architectures. To address these issues, we propose a number of methods: a multi-channel Data-Fusion Module (DFM), Neural Adaptive Patch (NAP) augmentation algorithm and re-weight class balancing (implemented in our PHR-CB experimental setup). We integrated these techniques into our novel Fusion Adaptive Patch Network (FAPNET). Our dataset is the Sentinel-1 SAR microwave signal, used in the Microsoft Artificial Intelligence for Earth competition, so that we can compare our results with the top scores in the competition. In order to validate our approach, we designed four experimental setups and in each setup, we compared our results with the popular image segmentation models UNET, VNET, DNCNN, UNET++, U2NET, ATTUNET, FPN and LINKNET. The comparisons demonstrate that our PHR-CB setup, with class balance, generates the best performance for all models in general and our FAPNET approach outperforms relative works. FAPNET successfully detected the salient features from the satellite images. FAPNET with a MeanIoU score of 87.06% outperforms the state-of-the-art UNET, which has a score of 79.54%. In addition, FAPNET has a shorter training time than other models, comparable to that of UNET (6.77 min for 5 epochs). Qualitative analysis also reveals that our FAPNET model successfully distinguishes micro waterbodies better than existing models. FAPNET is more robust to low lighting, cloud and weather fluctuations and can also be used in RGB images. Our proposed method is lightweight, computationally inexpensive, robust and simple to deploy in industrial applications. Our research findings show that flood-water mapping is more accurate when using SAR signals than RGB images. Our FAPNET architecture, having less parameters than UNET, can distinguish micro waterbodies accurately with shorter training time.

1. Introduction

Flooding often results in devastating consequences. Although societies have learnt over time how to create flood-prevention infrastructure, many areas continue to be affected by these disasters, costing property damage and lives. Climate change-induced flooding, whether from rising sea levels, extreme weather events, or shifting tides, is likely to affect nearly 4.8 million Canadians who live within 10 kilometers of the eastern or western coastline in the near future, according to a report published by the University of Waterloo [1]. Not only in Canada but also all around the globe, flooding is becoming increasingly common and can be devastating. Promisingly, studies show that it is possible to predict how severe the flooding may be based on the information available [2]. A combination of computer vision and machine learning techniques can be used to predict the occurrence of flooding, so that residents can evacuate [3].
Water level change and flood detection have traditionally been analyzed using imagery from passive sensors, such as optical images, applying approaches like the Normalized Difference Water Index (NDWI) and other water indices [4]. There are a variety of ways to combine multiple frequency bands to create signatures that are sensitive in detecting waterbodies [5]. Once an index has been calculated, it is possible to apply thresholds to detect water changes and flooding throughout a region. However, their use is confined to a specific time period and specific weather conditions, making them less ideal for rapid reaction in dynamic scenarios. As a result, Synthetic Aperture Radar (SAR) imagery has become a popular satellite data modality used for flood-water detection [6]. Because of its active sensing capability and specific frequency ranges, SAR images can be acquired at any time of the day despite cloud cover. As long as the image features can be extracted, high-resolution imaging is not always necessary.
Nevertheless, radar images are known to be impacted by polarized waves in the signal [7]. Our approach is to integrate the results of several polarization methods in order to provide a complete understanding of the ground surface, while a natural image offers three channels (red, green and blue). Polarization offers four combinations to analyze the separate channels: Vertical Transmission and Vertical Reception (VV), Vertical Transmission and Horizontal Reception (VH), Horizontal Transmission and Vertical Reception (HV) and Horizontal Transmission and Horizontal Reception (HH), where each channel provides unique characteristics. We may use one of these signals to detect flooding on individual chips but combining all the signals in semantic modeling will obtain accurate waterbody detection for flood mapping [8]. Supplementary elevation data (NASADEM) is a significant feature leading to higher accuracy for deep learning model training. Microwave signals are not suitable in conjunction with basic computer vision techniques such as indexing and thresholding.
In addition, GeoTiff radar signal preprocessing techniques have advanced rapidly in recent years. Radar imaging systems are often installed on aeroplanes or satellites with side-looking geometry. In radar imaging, shadow is a common geometric distortion that occurs in mountainous environments with high topographic variation. With the knowledge of topographic variation, it is feasible to extract waterbodies from shadows since the Digital Elevation Model (DEM) elevation values of most waterbodies tend to be the same [9]. For this reason, DEM was obtained from the corresponding coordinates of each polarized signal. Since microwave signals have a wide range of bands described as hyperspectral channels, data preprocessing becomes challenging and introduces additio rhead [10]. Deep learning models cannot directly manage a large number of channels. Hence, in this paper, we demonstrate that this issue can be addressed by combining (1) semantic modeling for selecting relevant channels (bands) and (2) efficient model design.
To obtain a high-level understanding of an image, semantic segmentation assigns each image pixel with a class label, which is known as pixel-wise classification [11]. Predefined features are frequently used in traditional machine learning applications for pixel-wise categorization [12]. A number of network architectures, including DenseNet [13], ResNet [14], EfficientNet [15] and VGG [16], have been introduced. These models are trained on different large datasets (e.g., imagenet [17]), which can be used as the encoder backbone in an image segmentation architecture to yield high segmentation accuracy. In recent years, there has been an increasing use of Deep Convolutional Neural Network (DCNN) models for waterbody identification [18,19,20,21,22]. According to the results and discussion [23], DCNN-based models are capable of extracting distinct spatial feature representations better than water index-based computer vision methods. In particular, FSSS has been shown to be the most effective technique for identifying water surfaces using satellite imagery or SAR.
Large intraclass differences and lower sensor resolution make it more difficult to identify objects in satellite images using image segmentation methods [24]. Recent advances include multimodality [25], hierarchical feature representations [26] and fusion techniques [25,27,28]. Three Convolutional Neural Network models are combined using MonteCarlo dropout uncertainty maps [29] to go beyond traditional weight averaging for land-cover mapping segmentation in urban zones. Kampffmeyer et al. [29] used a digital surface model established by Sun and Wang [30] to enhance the Fully Connected Network (FCN) to generate segmentation outcomes of Very-High-Resolution (VHR) remote sensed images.
A recent approach for flood-water mapping was proposed by Jiang et al. [31]. The authors used Sentinel-1 imagery, building their technique upon Attention-UNet3+ and using the Poyang Lake Region as a case study. They combined the feature maps from different scales to improve the accuracy of narrow waterbody detection. The authors claimed that by using a spatial attention module, false alarms caused by speckle noise and water shadows are reduced. However, the model requires a large number of parameters, which creates high computational overhead, with expensive GPU consumption during training. Instead, for narrow and tiny surface water detection, we use the NAP augmentation module and DFM during data preprocessing. We are able to obtain better results by adding external features provided by the Microsoft Planetary Data and DEM to reduce the number of shadow signals from the radar imagery. Another approach to detecting flood-water from RGB images was proposed by Sarp et al. [32], who used Mask-R-CNN. However, the authors only focused on detecting flood-water from roadways, which is for vehicle routing control and traffic management, in the smart autonomous vehicle application domain. To train their model, the authors used hand-crafted items. It is worth pointing out that their dataset is biased toward one class (background). Their evaluation was based on the F1 score, which is not guaranteed to be consistent with the MeanIoU score. Based on our study and the work of other researchers in the literature, MeanIoU is a reliable metric for segmentation performance evaluation because it calculates the average of each detected pixel, providing equal importance to each class. For this reason, we use the MeanIoU score for performance evaluation and comparison in our work.
In radar backscatter, there are similarities between flooded and unflooded conditions. Therefore, additional features are needed to boost the performance of the waterbody detection accuracy. Zhao et al. proposed Exclusion map (Ex-map), derived from Sentinel-I time series information, which aims to delineate the areas, where radar scatters do not detect flood-water [33]. By introducing Ex-map, the classification accuracy is increased significantly. Nevertheless, the lack of training data adversely affected the detection accuracy. To address this problem, a transfer learning-based flood detection technique, using RGB images, was proposed by Sazara et al. [34]. The authors suggested feature extraction techniques such as Local Binary Patterns (LBP), Histogram of Oriented Gradients (HOG) and pre-trained deep neural network (VGG-16), followed by water surface classification. Trained logistic regression, k-nearest neighbours and decision tree classifiers were used. However, optical RGB images contain less features than an SAR signal. SAR provides more information capable of wide area monitoring even at night as well as in worse weather conditions. Due to a larger spectrum of data, SAR signals require special preprocessing steps in order to benefit from the complex SAR features and achieve more accurate flood-water detection.
Considering the drawbacks of the current methods and based on our analysis, we propose a novel approach for a flood-water detection and monitoring system that embeds multi-channel fusion, unit kernel-based patch creation and a FAPNET model. Our method includes data fusion to handle multiple image resolutions and normalization of Sentinel-1 VV, VH and elevation data from the NASADEM bands. We also created a new NAP augmentation module to keep the high-dimensional representations prominent. NAP augmentation reduces patch generation data loss while improving model performance. By comparing with related methods [8], our experiments have shown better performance than existing deep learning approaches.
One novel idea introduced in this paper is the successful design and validation of FAPNET, which integrates the NAP augmentation algorithm and DFM. FAPNET can process both radar signals and optical images efficiently. However, the focus of this paper is on satellite SAR data analyzing invisible hyperspectral bands which provide more feature information. Applying FAPNET on the visible RGB bands for other applications can be studied in future work. NAP is a new and unique concept, which is different from the traditional augmentation method. It optimizes memory usage, which is important for industrial applications, where computing resources can be a bottleneck. We validated the better performance of our proposed model FAPNET by comparing it with related works, using qualitative, quantitative and memory usage analysis. Our contributions are summarized below;
  • In order to maximize the effectiveness of the learned weights, we introduce a multi-channel DFM using VV, VH and elevation data from the NASA Digital Elevation Model (NASADEM), incorporating Feature Fusion, normalization and end-to-end masking.
  • Inspired by the CNN architectures, we introduce an effective augmentation technique by creating an NAP augmentation module to extract multi-receptive features as patches using unit kernel convolution at multi-scale levels, which helps to achieve an enriched semantic representation and helps the model to learn faster.
  • We confirm the efficacy of our method by designing four experimental setups: CFR, CFR-CB, PHR and PHR-CB, which build a progressive analytic framework for researchers to investigate semantic segmentation in SAR images and conduct comparisons with state-of-the-art methods.
  • We propose FAPNET, which is a lightweight, memory-time efficient and high performance model, as demonstrated in our quantitative and qualitative analysis.
The rest of this article is structured as follows: Section 2 explains data preparation, feature mapping and data augmentation. Details of our proposed Fusion Adaptive Patch Network (FAPNET) model and NAP augmentation algorithm are explained in Section 3. Section 4 provides quantitative, qualitative and memory-time performance assessments and Section 5 discusses our findings. Conclusion and future work are given in Section 6.

2. Study Area and Dataset

We used the “Map flood-water from Radar Imagery” competition dataset, which is hosted by Microsoft AI for Earth [8]. The dataset consists of Sentinel-1 images, masks and a CSV file containing metadata such as city and year. Sentinel-1 images and masks were collected from different regions of the world between 2016 and 2020. Sentinel-1 is able to send and receive signals in both horizontal and vertical polarizations since it is a phase-preserving dual polarisation SAR device. The dataset contains 542 chips as features and corresponding flood-water mapping as masks. Based on how radar microwave frequencies are dispatched and received, a single chip characterizes two bands or images (VV, VH). Each image is saved as a 512 × 512 resolution GeoTIFF file and each image pixel value denotes the amount of energy, expressed in decibels, reflected back to the satellite from the ground. Chips from 13 flood occurrences are included in the dataset and each flood event contains about 15 to 69 chips (30 to 138 images), with half of the events comprised of fewer than 32 chips (64 images). Table 1 shows a summary of the dataset divided by locations.

2.1. Feature Mapping and Discrepancy in Masking

According to our observations in Figure 1, it is obvious that the dataset has some issues, i.e., surface water information is missing in the features (VV, VH, DEM) but available in the corresponding masks. When visualizing features with their respective masks, it is noticeable that the extracted features indicate fewer water pixel values than the mask. Figure 1 illustrates a few examples. Figure 1c is an example of a consistent feature map and mask from the dataset. However, in Figure 1a,b, we can see discrepancies between the features and masks; in particular, Figure 1a shows significant differences. Moreover, several feature values in Figure 1b do not correspond to the water information correctly and the right-hand-side white patch of pixels in the VV and VH images reflects missing values, which might be due to human error or distortion caused by misinterpreted radar frequency.
To solve this problem, a perfect mask (label) creation is required, and we need a unique technique to minimize the errors. Consequently, we introduce a novel data augmentation technique called the NAP augmentation algorithm, which will be discussed briefly in Section 3.2. Note that the NAP augmentation algorithm is designed not only to minimize the error of mislabeling but also to boost the performance. In addition, many surface characteristics, such as elevation, terrain and ground property, can cause water to disperse or accumulate in one place; we tried several techniques and different experimental setups to evaluate the model performance. Our finding shows that, given the different water level changes in various regions, accurate feature mapping dictates the success of water pixel or flood prediction.

2.2. Data Augmentation

Albumentations is a fast and flexible data augmentation tool commonly used in industrial projects, especially in open source projects. Most image transform operations and certain domain or task-specific analyses, such as weather change predictions for autonomous vehicle perception modeling, are supported by Albumentations. Despite other libraries applying bilinear interpolation, its resizing and rotating process can adversely affect the task outcome because such data augmentation approaches might modify the mask value of the associated augmented feature [35]. Buslaev et al. [35], on the other hand, suggest employing regression unit tests to validate the augmented mask. Moreover, they reported state-of-the-art results for the segmentation task on the Inria Aerial Image Labeling dataset [36] after using several augmentation methods. A regression unit test is included in the Albumentations tool. We will describe our methodology in the next Section.

3. Methodology

In Section 2, we explained the study area and other important information regarding the dataset. The Microsoft AI for Earth Competition identifies VV, VH and DEM as important features for flood-water detection and therefore we use the same input features in our experiments [8]. Competition also highlights the significance of both signals and the need for DEM data in order to achieve improved accuracy, particularly when training deep learning models. Natural terrain and persistent water sources in a region’s environment might further aid a deep learning model’s ability to identify flooding. When processing the raw images, we included elevation information from NASA’s Digital Elevation Model (NASADEM). As illustrated in Figure 2, data is passed through the Albumentations tool [35] for augmentation and is further split into training (80%), validation (10%) and testing (10%) samples. In order to measure the model’s competence while tuning its hyperparameters, we used a validation dataset, which was a subset of the data used during training. In order to objectively evaluate how well a model fits the training data, a separate test dataset was deployed. In this study, we compared the outcomes based on the test dataset to provide a comprehensive evaluation.

3.1. Overall Strategy

In this work, we introduce the FAPNET model, which comprises two new techniques: the NAP augmentation method and the DFM. We fuse three single channel signals, VV, VH and DEM, in the DFM. Then we normalize the Feature Fusion as shown in Figure 2. In order to evenly distribute the classes in the dataset, we swap the unlabeled pixel values of the mask into the background. Next, we apply the NAP augmentation algorithm, which will be explained in Section 3.2. In order to evaluate our proposed FAPNET model, we designed four distinct experimental setups CFR, CFR-CB, PHR and PHR-CB and tested across a wide range of water dynamic scenarios. We compared FAPNET with existing segmentation models—UNET, VNET, DNCNN, UNET++, U2NET, ATTUNET, FPN and LINKNET. CFR and CFR-CB take only 512 × 512 input before running the DFM module. PHR and PHR-CB take 256 × 256 input before running the DFM and NAP augmentation algorithm. The same parameter settings and preprocessing techniques are used in all experiments. The four distinct experimental setups serve to validate and compare different models’ performances in various real-world contexts. The assessment metrics are MeanIoU and focal loss, which will be further discussed in the Quantitative Assessment section. Our FAPNET model is inspired by the encoder−decoder-based structure of UNET, but the number of hidden layers and convolution layers is different; FAPNET analyzes more deeply into the hyperparameter settings as described in Section 3.4. We validated our FAPNET model performance using quantitative, qualitative and memory-time efficiency analyses.

3.2. Proposed NAP Augmentation Algorithm

The majority of the augmentation techniques use a bilinear interpolation method that can affect data analysis and evaluation outcomes. To address this issue, Albumentations [35] recommends a regression unit test for validation. However, this does not guarantee the correct transformation of pixel values when arbitrarily rotating or flipping images. Inspired by the CNN architecture [37], our proposed NAP augmentation module makes use of a unit kernel and pixel-wise convolution that does not transform the pixel values and instead extracts the pixel values as patch images.
We use the NAP augmentation module to extract patches of defined size from each image in the dataset. Current patch libraries that we have explored so far lack the capability to overlap extracted features and do not ensure class balance. Some tools distort feature values while extracting image patches. In our NAP augmentation module, we deploy the CNN architecture concept and preserve additional parameters that make it possible to extract overlapping features and maintain class balance.
Note that any factor like the input shape, the kernel, zero padding, dilation and the strides used can affect the output of a convolutional layer. However, it is not easy to determine the relationship between these variables [38]. We use Equation (1) to illustrate the relationship between these variables as suggested in [38],
o u t = y + 2 p k ( k 1 ) ( d 1 ) s + 1
Here, y is the input shape, p is padding, k is the kernel size, d is the dilation that increases the respective field of output without changing the kernel size and s indicates how much the kernel moves vertically and horizontally through the input matrix.
Changing the patch size in Convolutional Neural Network (CNN) can offer robust performance regardless of dataset variation [39]. We propose the NAP augmentation module, with ideas inspired by the CNN architecture (Equations (2)–(4)). Our padding ( p ) value is 0 since we generate patches from a given image by applying pixel-wise convolution. The dilation ( d ) value can be utilized to simplify the CNN-based architecture. On the other hand, ref. [40] proposed a method for training CNN-based architecture that uses dilation values, without considering coverage or resolution loss. However, since our NAP augmentation module performs pixel-wise convolution to extract patches from a given image before training the CNN-based architecture, the patches could be distorted. Therefore, our dilation ( d ) value is set to 0. When we substitute all these numbers into Equation (1), we can see that our Equations (2) and (3) are simplified versions of Equation (1). We introduce three new equations derived from Equation (1) [38] to accomplish the NAP algorithm. The NAP augmentation module is shown at the bottom right corner of Figure 2. A full procedure is given in the Algorithm 1.
p h = y k s + 1
p l = y k s + 1
N = p h × p l
Algorithm 1:NAP Algorithm
Sensors 22 08245 i001
Here, p h represents the number of times the unit kernel moves horizontally and  p l represents the number of times the unit kernel moves vertically, dictated by the image dimension, patch size and stride. N represents the number of patches generated from a given image.

3.3. Proposed FAPNET Model

The architecture of our proposed FAPNET model is illustrated in Figure 3. The three input channels consist of the VV, VH and DEM. The DFM processes these input channels and identical-sized feature maps are generated and concatenated before passing as input into our encoder–decoder network for pixel-level labeling. In our encoder–decoder architecture, we do not utilize any pre-trained model weight. However, our NAP augmentation module, as stated in Section 3.2, enhances our model’s performance. The entire network configuration is described in Table 2 and Table 3, using various experimental setups, which will be discussed in Section 4. Tables include the kernel width, kernel height and number of kernels for each convolutional layer, as well as the output sizes of the feature maps. A summary of our model algorithm is given in Algorithm 2.
Algorithm 2:Proposed FAPNET Model Algorithm
Sensors 22 08245 i002
The model uses two blocks: a d o w n b l o c k and an u p b l o c k to down-sample and up-sample. The first layer, d o w n b l o c k -1 and u p b l o c k -1, is made up of 2-kernel ( 3 × 3 ) convolutional networks with the same padding size, followed by a ReLU activation function and a dropout layer. D o w n b l o c k -2 through d o w n b l o c k -6 have a similar structure to d o w n b l o c k -1, but are followed by a 2 × 2 max-pooling layer to downsize the feature height and width by half. U p b l o c k -1 is constructed similar to d o w n b l o c k -1, except u p b l o c k -1 is followed by a 2 × 2 convolutional layer with same padding size and SoftMax activation. In addition, each layer’s d o w n b l o c k has a skip connection that connects it to its equivalent u p b l o c k .
Each convolutional layer of a single d o w n b l o c k contains the same number of kernels and is multiplied by two for the following d o w n b l o c k s up to d o w n b l o c k -4. In d o w n b l o c k -5 and d o w n b l o c k -6, the second convolutional layer kernel numbers are decreased by one-fourth and half of d o w n b l o c k -4, respectively. On the other hand, the number of kernels in u p b l o c k -6 to u p b l o c k -4 is similar, while the number of kernels in u p b l o c k -3 to u p b l o c k -1 is divided by 2 with regard to the preceding u p b l o c k , as shown in Table 2. Our proposed FAPNET m i d d l e b l o c k is made up of a max-pooling layer having two 3 × 3 convolutional networks with ReLU activation function and a dropout layer, all of which are connected by 2 × 2 convolutional transpose with the same padding size.

3.4. Hyperparameter Tuning

Hyperparameters have the potential to directly influence the training algorithm’s behavior. Choosing optimal hyperparameters has a significant impact on the performance of the trained model. Traditionally, the optimal hyperparameter values for a given dataset were determined manually. Researchers usually lean on their prior experience in training machine learning algorithms to choose these parameters. The disadvantage is that the optimal setting for hyperparameters used to solve one problem may not be optimal for a different problem, as these values vary between datasets and even image categories. Consequently, it is challenging to establish the hyperparameter values only based on past experience [41].
To improve our model dropout and number of kernels in each layer, we utilize the auto Keras framework [42]. The framework uses Bayesian optimization to guide through the search space by identifying the most promising operations each time, resulting in an efficient neural architecture search for hyperparameter tuning. It also utilizes an edit-distance neural network kernel to determine the number of operations needed to transfer from one neural network to another [42]. After tuning, we obtain the optimal number of kernels in each layer, as shown in Table 2 and Table 3, where in each conv2d layer ( C o n v 2 D 1 and C o n v 2 D 2 column), inside the parenthesis it indicates the feature’s height, width and the number of the kernels, respectively.
When gradient descent hits a cost surface minimum, parameter values might fluctuate around the minimum. To counteract this, gradient calculation can be optimized by reducing the learning rate [43]. To resolve the issue, we apply a learning rate scheduler that reduces the learning rate after certain epochs.

3.5. Loss Function

Focal Loss (FL) [44] is a BinaryCross-Entropy variation. It reduces the importance of easy examples, allowing the model to concentrate more on learning difficult instances. It performs well in circumstances where the classes are very imbalanced. By using Equation (5), we can calculate FL.
F L ( y , y ) = α ( 1 y ) γ y log ( y ) ( 1 α ) y γ ( 1 y ) log ( 1 y )
where y is the ground truth and y is the model prediction. The α value is 0.25 and can be treated as a hyperparameter. γ can also be treated as a parameter, but as per [44], setting its value to 2 provides the best result.

3.6. Evaluation Metrics

Since we perform a binary segmentation, there are two possible outcomes: positive or negative, which correspond to water or background area, respectively. True Positive (TP) is a segmentation result that shows water is identified correctly. True Negative (TN) indicates that background pixels are classified correctly. False Positive (FP) refers to background incorrectly identified as water. Water pixels are classified incorrectly as background if False Negative (FN) occurs. Several measures can be used to assess segmentation performance. In general, we want to know the percentage of pixels that have been correctly identified, regardless of their class. Precision, recall, F1 [45], the Mean Intersection Over Union (MeanIoU) [46] of the water class and Dice Coefficient [47] scores are often used in the literature.
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
For the segmentation evaluation, the MeanIoU [46] score is a commonly used performance metric. The MeanIoU [46] measure is given in Equation (9).
I o U = T P T P + F P + F N
The Dice Coefficient [47] is a broadly used indicator in the computer vision domain to determine the similarity of two images [48]. It has evolved into a loss function called Dice Loss [47].
D C ( y , y ) = 2 × | y y | + 1 y + y + 1
where y is the ground truth and y is the model prediction. In the numerator and denominator, 1 is added so that it does not become undefined if y = y = 0

4. Performance Evaluation

We used three different evaluating metrics—quantitative, qualitative and memory-time efficiency analysis—to verify the effectiveness of our NAP augmentation technique and DFM, integrated in the proposed FAPNET architecture. We will describe our four experimental designs in this section, along with the results that they produced. Even though we output the training, test and validation results, we focus on the test dataset for performance evaluation because the test data is completely unseen during training. FAPNET is compared with other segmentation models UNET, VNET, DNCNN, UNET++, U2NET, ATTUNET, FPN and LINKNET. Based on the evaluating metrics, quantitative, qualitative and memory-time efficiency analyses, FAPNET consistently outperforms other models. Since the parameter settings and preprocessing steps are the same across all experimental setups (CFR, CFR-CB, PHR, PHR-CB), the comparison is fair. The quantitative analysis described in Section 4.1 includes an explanation of the experimental setups. In addition to comparing our outcomes with those of the competitors (Section 4.9), we also analyze the significance of the DFM in Section 4.8.

4.1. Quantitative Analysis

There are a number of evaluation metrics as described in Section 3.6 above. Since MeanIoU combines precision and recall statistics into a compact score [46], it has become increasingly used in the literatures over the past decade. Therefore, in our experiments, we will focus on using MeanIoU to analyze the test dataset results, which are generated from four setups: CFR, CFR-CB, PHR and PHR-CB. Figure 4 shows an overview of the four experimental workflows. We focus on the MeanIoU score under the testing dataset column because it reflects the outcome after model training.

4.2. CFR

In this setup, the data is passed through the DFM, similar to the state-of-the-art methods mentioned in the “Map flood-water from Radar Imagery” competition. All images/chips have the same resolution ( 512 × 512 ) . We use data augmentation and data normalization. Pixel information is given in Table 4, which will be required in Section 4.3, CFR-CB experiment and the results of the CFR experiments are shown in Table 5. FPN is created to comprehend LiDAR point cloud [49] and it can be seen that the FPN [49] model is the second best. Our FAPNET performs best overall and has the best MeanIoU of 0.8544 in the testing dataset.

4.3. CFR-CB

In the CFR setup, results may be affected by class imablance, which can lead to unanticipated inaccuracy or significant adverse effects in data analysis, particularly during the classification tasks [50]. From Table 4, we can see that background (no water) populates 86 % in the testing dataset and in our CFR setup above, we did not address the data imbalance problem. In this CFR-CB experiment, we introduce a re-weighting technique to adjust the imbalance ratio by inversely multiplying our mask with the class weights shown in Table 4. Loss balancing is also obtained as a result of this re-weighting method [51].
The results of the CFR-CB experiment are shown in Table 6. After balancing the dataset by weight, most models’ scores rise by at least 1 % to 6 % in general compared to Table 5. However, for the DNCNN [52] and ATTUNET [53] models, the opposite occurs. their model architecture may be sensitive to the class balance weight. FPN remains the second best. However, LINKNET has the biggest improvement, achieving a MeanIoU of 0.8638 from 0.8021, overtaking UNET as the third best. On the other hand, FAPNET surpasses other models in terms of the MeanIoU on the testing dataset with a score of 0.8544.

4.4. PHR

In our previous two experiments, we use the classical data augmentation technique and class balance in the CFR-CB setup. However, in this PHR experiment, we use the NAP augmentation module described in Section 4.5 without class balancing. We generate image patches with patch size ( k ) 256 and stride ( s ) value 64. As mentioned before, the image shape in the dataset is 512 × 512 and by passing a single image through the NAP augmentation module, we get 25 images of dimension 256 × 256 .
Table 7 shows that nearly all models (except U2NET) give inferior scores compared to Table 5 and Table 6. Note that in this setup, we generate patches from the dataset, producing around 10K patches, which also creates many patches that do not have any water pixels, leading to large variations between classes. As a result, the models are biased towards the non-water class, affecting the outcome accuracy. Our FAPNET appears more sensitive to this class imbalance compared to other models. In order not to introduce bias in the evaluation, we design PHR-CB, taking class balance into consideration.

4.5. PHR-CB

We can see from the PHR experimental setup that without class balancing, the outcome can be biased in favor of models that are “data hungry” or have a high number of parameters. In the PHR experiment, we generate approximately 10K patches from 433 training images, but it leads to a data imbalance problem because the NAP augmentation approach generates many patches that do not contain any water or have a low proportion of water, due to the image content of the “Map flood-water from Radar Imagery” competition dataset. We solve this issue by rewriting Equation (4) as shown in Equation (11) for this PHR-CB experiment. In Equation (11), we set α to 0.20 , adjusting the water and non-water proportion back to the proper class distribution.
The number of alleviated patches after using the NAP augmentation approach is denoted by α in Equation (11).
N = p h × p l α
Table 8 shows that overall our proposed FAPNET model outperforms other MeanIoU scores in the test dataset. It even shows better results than its performance in other experiments, including CFR-CB, which is also class balance. This is attributed to changing the NAP augmentation module Equations (4)–(11), which reduces the high class variation. It is worth pointing out that compared to all other models and all experimental setups, FAPNET provides the best MeanIoU scores of 0.9514, 0.8960 and 0.8706 on training, validating and testing samples, respectively, in the PHR-CB experiment.
Based on the quantitative analysis, we can conclude that our model is sensitive towards class imbalance and the PHR-CB experimental setup, which incorporates the NAP augmentation and class-balance technique, is most suitable and unbiased in terms of addressing the class imbalance for all models. In Figure 5, we compare the training, validation and testing dataset MeanIoU score for the experiments across the models.

4.6. Qualitative Analysis

We present in this section our qualitative analysis, which also demonstrates that our approach outperforms other models. From our testing dataset, we randomly selected three predictions, which are shown in Figure 6, where the first four rows represent VV, VH, DEM and Ground Truth (GR), respectively. The remaining rows show model predictions. Three examples are given and the results from the four experimental setups are shown under each example. In the CFR experiment, it is noticeable that U2NET [57], DNCNN [52], ATTUNET [53], UNET++ [56] and VNET [55] cannot correctly predict water pixels in all three examples. In contrast, our FAPNET, UNET [54], FPN [49] and LINKNET [58] can accurately predict more water pixels than other models. In the second example, we can observe that our model predicts water pixels more accurately than other models. UNET [54], FPN [49] and LINKNET [58] are unable to predict the top left corner of the expected mask and are less accurate than our model (as shown in the final row, fifth column). Another example of this behaviour can be seen in the last row, ninth column (top center and right side of the predicted mask). We can also analyze each experimental setup in Figure 6. Overall, our proposed FAPNET outperforms all other models in the CFR experimental setup. The majority of the models perform well in the CFR-CB experiment, capturing more water distributions.
In the PHR experiment, we observe that FPN [49], UNET [54], LINKNET [58] and our proposed FAPNET models (all performed well in the CFR and CFR-CB experiments) tend to diverge from the ground truth. The dataset size in this experiment was approximately 10 k and there were many patches without water pixels, as described in Section 4.5, which is attributed to the decreased model performance due to class imbalance. In contrast, the prediction of U2NET [57] improves, because it can extract more features from larger datasets. In addition, a closer look at the third example shows that, with the exception of FPN [49], all other model predictions are lower in this experiment. We can conclude from this finding that highly imbalanced data affect accurate model prediction. Increasing the number of training samples can also introduce high class variance. That is the reason we introduce the PHR-CB experimental setup.
PHR-CB is the most successful experimental design since every model prediction increased significantly in terms of qualitative analysis. As can be seen in Figure 6, the predictions of U2NET [57], DNCNN [52], ATTUNET [53], UNET++ [56] and VNET [55] improve substantially in each example. However, note that some models have a tendency to predict noise and our FAPNET model predicts a certain amount of noise in the first and third examples, but the noise is considerably less than other existing models. In Figure 6, the first example, we can observe that the FAPNET model accurately predicts water distribution that other models overlook. In Figure 6, the second example, our model’s prediction is the most accurate among all models, and in the third example, FAPNET predicts almost all of the water pixels with a very low level of noise. Some small waterbody pixels are not properly identified in any of the models’ predictions, whereas FAPNET detects more small water areas as well as larger area, especially in the PHR-CB experiment.
Overall, the results demonstrate that the PHR-CB and CFR-CB experiments, with class balancing, are the best setup for qualitative performance evaluation. The PHR-CB experiment is the best for quantitative evaluation as all models increase their performances in general, with our FAPNET model outperforming others.

4.7. Memory-Time Efficiency Analysis

In this Section, we compare the model’s GPU, RAM and memory usage, as well as the run-time for different batch sizes and epochs. There are trade-offs between accuracy, memory usage and time performance when evaluating a model. A model’s number of parameters may also indicate whether it is lightweight. When developing the Refactor Equation (12), we chose to deploy U2NET [57] parameters because of its large parameter values as the numerator. If the refactor value is lower than the others, we know that the model is lighter than the others.
r e f a c t o r = p a r a m e t e r s h i g h e s t _ p a r a m e t e r s
From the refactor value, we can see that our proposed model FAPNET contains more than three times as many parameters as UNET [54], as shown in Table 9. Yet its performance in terms of memory usage and training time is comparable to UNET [54]. Models with huge amount of parameters and complicated architectures tend to extract more features, e.g., U2NET [57] and ATTUNET [53], which increases training time and resource usage, as shown in Table 9. In addition, models with complicated architectures are data-hungry, as observed in the PHR experiment, where the U2NET [57], VNET [55] and UNET++ [56] models perform well as the dataset is six times larger. On the other hand, models with simple architectures, such as UNET [54] and our proposed model FAPNET, can perform well in all experiments in terms of time performance, memory usage and accuracy. Overall, among the models with a simple design, our model outperforms UNET [54], as demonstrated by the quantitative and qualitative evaluations.

4.8. Significance of DFM Module

From the above experimental analysis, it can be seen that our FAPNET model yields the best results in the PHR-CB experimental setup when using the DFM module. However, to assess the importance of VV, VH and DEM features independently, we need to conduct a new experiment using each feature as input, followed by a result comparison using qualitative, quantitative and memory-time efficiency analyses.
In this section, we discuss the importance of the DFM module by conducting more experiments, in which we employed different combinations of model input (VV, VH, DEM and DFM) and compared the prediction accuracy. We also checked the significance of each feature of the DFM module and discovered which feature has more influence over flood-water prediction. We used the FAPNET model and the PHR-CB experimental setup to compare the effectiveness of individual input features based on qualitative, quantitative and memory-time efficiency analyses. Table 10 shows the quantitative analysis for each input feature, where DFM outperforms others with a MeanIoU score of 0.8706, while the VV, VH and DEM features obtain MeanIoU scores of 0.5405, 0.6063 and 0.5404, respectively.
The results show that VH has more influence than VV for flood-water prediction, while DEM only provides high-level features. Figure 7 shows the predictions from the VV and VH features are too noisy, while the DEM prediction is almost null. However, the DFM prediction is less noisy and almost similar to the ground truth. This is because we combine scaling with normalization in the DFM module, which enhances the model performance. We also performed memory-time efficiency analysis. Table 11 shows that DFM takes longer time to train the model than other feature models. While the CPU consumption is similar, the GPU consumption in DFM is higher than other feature models. Overall, FAPNET provides the best performance in quantitative and qualitative analysis, but has a longer training time and memory consumption.

4.9. Comparison with Competition Results

The best-performing model in the Microsoft AI on Earth Competition achieved a MeanIoU score of 0.8094, showing a significant increase over the benchmark result (0.4357). The second and third positions are slightly less than the top position—0.8072 and 0.8036, respectively. All suggested architectures are able to accurately recognize the larger water surface area. However, they all use segmentation techniques without taking into account the relevance of tiny water pixels (small salient features). Moreover, no data preparation or data augmentation methods were used on top of the UNET-based benchmark solution. As polarized SAR data make available signals from a wide variety of landscapes and plant types, all winning approaches from the competition utilized the Planetary Computer STAC API to incorporate supplementary data from the NASADEM and/or the European Commission’s Joint Research Centre (JRC) global surface water data in their models, leading to a better understanding of the geographic area’s natural topography. We designed a new FAPNET architecture for better detection of tiny water pixels, which are often missed by existing models. Our DFM is used for noise suppression and NAP augmentation is used to boost both quantitative and qualitative performance. Figure 8 shows that FAPNET performs better than the top scores in the competition.

5. Discussion

Our motivation to introduce the FAPNET architecture came from the UNET [54] and our model contains three times as many parameters as UNET [54]. Though the FAPNET refactor 0.13 is almost the same as other existing models, it takes less time for training, as reported in Table 9. Hyperparameter tweaking also makes our model more versatile in terms of extracting more salient features. Table 9 demonstrates that, although having more parameters than UNET [54], our FAPNET model uses the same amount of time to train and outperforms other models with the same number of parameters, i.e., FPN [49], UNET++ [56] and LINKNET [58]. In general, the class balance in the CFR-CB experiment enhances model performance for all models, compared to the CFR experiment. This indicates that data preprocessing and class balance are essential factors for increasing model performance. However, we also found that data augmentation incorporated in our proposed NAP augmentation module has both pros and cons that differ between models and experimental setups. In the PHR experiment, our FAPNET model is underrated in comparison to other models. Yet, the class balance setup in the PHR-CB experiment gives it an advantage over other models and experiments, leading to the best MeanIoU scores in training, validating and testing datasets. Overall, combining the PHR-CB experimental setup with our proposed FAPNET model outperforms all other existing models in terms of quantitative, qualitative and memory-time efficiency performance.
For two reasons, training data collection has increasingly become a major concern. First, as machine learning is becoming more popular and more commonly used in industry development and academic research, new applications are emerging that require greater varieties of labeled data. Second, unlike traditional machine learning algorithms, deep learning algorithms explore complex feature representations, cutting costs on feature engineering, which often requires more labeled data [59]. The dataset we used in our experiments has some flaws, as stated in Section 2.1. Fixing these issues will likely affect the model performance but might produce better results. Model performance may be affected when shifting from traditional data augmentation strategies to cutting-edge techniques like feature transplanting, style transfer and GAN-generated augmented data. As discovered in our PHR-CB experiment, input shape and class balance can affect model performance. Hence, it is worthwhile to investigate a broader scope of possibilities, e.g., testing whether model performance is affected by varying the input shape. Class balance has a direct impact on loss calculation and it is an important aspect in deep learning model training. Note that adopting a custom loss function may influence the model’s performance. Additionally, we should be able to enrich our model by employing the hyperparameter tuning approach with regards to dropout and the number of kernels. Performance will likely be improved if we adjust other hyperparameters using various optimization methods.
UNET was initially invented to address medical image segmentation problems. Later, researchers applied the UNET architecture on pixel-level classification in numerous applications. Radar signals have complex hyperspectral characteristics beyond the visible spectrum, which often creates challenges in the analysis process. To the best of our knowledge, hyperspectral image processing using deep learning is still inadequately explored in the research community. We preprocessed hyperspectral images and introduced our FAPNET model, with more fine-tuned parameters than other existing models. The results outperformed related works, especially when combining with the NAP augmentation algorithm and DFM module. Experimental results show that FAPNET performs better than UNET with comparable training time. We believe that FAPNET can perform well on optical images in terms of quantitative, qualitative and memory-time efficiency analysis, because the RGB bands (long, medium and short wavelengths) are a part of the electromagnetic spectrum. Since the focus of this work is for flood mapping-related analysis based on the invisible band features of the spectrum, we will leave the optical images to future work when studying other applications.
Another idea for future work is to analyze multiple features simultaneously inside our DFM and increase the number of channels (e.g., 4–6 channels instead of 3 channels) to accommodate both the visible RGB and invisible radar signal bands. We will be able to compare the results of this neural net architecture with the outcome from transfer learning discussed above. For now, FAPNET outperforms other models for detecting waterbodies using the invisible bands as input.

6. Conclusions

In this paper, inspired by the widely adopted deep learning techniques in satellite remote sensing, we propose a variety of novel methods for extracting waterbody from satellite imagery. In comparison to prior models, our proposed FAPNET model utilizes fewer parameters while achieving superior qualitative and quantitative results. By introducing the DFM module and NAP augmentation algorithm in the data preprocessing stage, our FAPNET model delivers better performance and reduces computational overhead compared with related works. Our DFM combines and normalizes feature bands, while the NAP augmentation algorithm creates image patches in the required resolution without altering the pixel values. We compared our NAP algorithm with commonly used augmentation techniques (rotation, flip, etc.) and found that by adopting NAP, the model could identify even a tiny water surface. Our proposed FAPNET, which incorporates DFM and the NAP augmentation method, outperforms the state-of-the-art methods in the PHR-CB experimental configuration based on the MeanIoU score, memory consumption and training execution time. The tuning of the hyperparameters helps in choosing optimal parameter values. Qualitative evaluation shows that our model outperforms other data preprocessing methods and deep learning frameworks like UNET when it comes to detecting the presence of water. We anticipate that our model will be able to identify even the tiniest waterbody; however, this is balanced by the fact that it also forecasts a fair amount of background noise. The traditional UNET-based model structure works well in detecting big water regions. Nevertheless, we observed that an augmentation mechanism may assist deep learning models in recognizing tiny areas with less noise (encoder–decoder-based model). Transforming from older data augmentation methods to more modern techniques could further enhance model efficiency. We hope to explore more advanced techniques in future work, including feature transplanting, style transfer and generative adversarial network (GAN). In our PHR-CB experiments, we discovered that the model’s performance was affected by the diverse input feature shapes and class re-weighting. We will look into a new model layer design that can accommodate a wide variety of input shapes. For class re-weighting, we will look into new loss functions.

Author Contributions

Conceptualization, M.S.I.; methodology, M.S.I.; software, M.S.I.; validation, M.S.I.; formal analysis, M.S.I.; investigation, Z.W.; resources, Z.W.; data curation, M.S.I.; writing—original draft preparation, M.S.I.; writing—review and editing, X.S.; visualization, M.S.I.; supervision, I.C.; project administration, I.C.; funding acquisition, I.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSERC grant number CRDPJ 543428-19.

Informed Consent Statement

Not applicable.

Data Availability Statement

Microsoft AI for Earth [8].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Intact Center on Climate Adaption; University of Waterloo. Rising Seas & Shifting Sands; Technical Report. 2021. Available online: https://bit.ly/3akf0GX (accessed on 2 May 2022).
  2. Tang, D.; Wang, F.; Xiang, Y.; You, H.; Kang, W. Automatic Water Detection Method in Flooding Area for GF-3 Single-Polarization Data. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 5266–5269. [Google Scholar] [CrossRef]
  3. Kwak, Y.; Yun, S.h.; Iwami, Y. A new approach for rapid urban flood mapping using ALOS-2/PALSAR-2 in 2015 Kinu River Flood, Japan. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 1880–1883. [Google Scholar] [CrossRef]
  4. McFeeters, S.K. The use of the Normalized Difference Water Index (NDWI) in the delineation of open water features. Int. J. Remote Sens. 1996, 17, 1425–1432. [Google Scholar] [CrossRef]
  5. Farhadi, H.; Esmaeily, A.; Najafzadeh, M. Flood monitoring by integration of Remote Sensing technique and Multi-Criteria Decision Making method. Comput. Geosci. 2022, 160, 105045. [Google Scholar] [CrossRef]
  6. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
  7. Turčinović, F.; Erny, M.; Bosiljevac, M. Polarization Sensitivity of Short-Range S-band Radar Sensing System. In Proceedings of the 2021 International Symposium ELMAR, Zadar, Croatia, 13–15 September 2021; pp. 127–130. [Google Scholar] [CrossRef]
  8. STAC Overflow: Map Floodwater from Radar Imagery. Available online: https://www.drivendata.org/competitions/81/detect-flood-water/page/386/ (accessed on 1 June 2022).
  9. Lachaise, M.; Schweisshelm, B.; Fritz, T. The New Tandem-X Change Dem: Specifications And Interferometric Processing. In Proceedings of the 2020 IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS), Santiago, Chile, 22–26 March 2020; pp. 646–651. [Google Scholar] [CrossRef]
  10. Kumar, M.S.; Keerthi, V.; Anjnai, R.N.; Sarma, M.M.; Bothale, V. Evalution of Machine Learning Methods for Hyperspectral Image Classification. In Proceedings of the 2020 IEEE India Geoscience and Remote Sensing Symposium (InGARSS), Virtual, 1–4 December 2020; pp. 225–228. [Google Scholar] [CrossRef]
  11. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  12. Guo, J.; Zhou, H.; Zhu, C. Cascaded classification of high resolution remote sensing images using multiple contexts. Inf. Sci. 2013, 221, 84–97. [Google Scholar] [CrossRef]
  13. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  15. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; Volume 97, pp. 6105–6114. [Google Scholar]
  16. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  17. ImageNet. Available online: https://www.image-net.org/ (accessed on 25 July 2022).
  18. Miao, Z.; Fu, K.; Sun, H.; Sun, X.; Yan, M. Automatic Water-Body Segmentation From High-Resolution Satellite Images via Deep Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 602–606. [Google Scholar] [CrossRef]
  19. Zhang, J.; Xing, M.; Sun, G.C.; Chen, J.; Li, M.; Hu, Y.; Bao, Z. Water Body Detection in High-Resolution SAR Images With Cascaded Fully-Convolutional Network and Variable Focal Loss. IEEE Trans. Geosci. Remote Sens. 2021, 59, 316–332. [Google Scholar] [CrossRef]
  20. Geng, J.; Fan, J.; Wang, H.; Ma, X.; Li, B.; Chen, F. High-Resolution SAR Image Classification via Deep Convolutional Autoencoders. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2351–2355. [Google Scholar] [CrossRef]
  21. Zhou, Y.; Wang, H.; Xu, F.; Jin, Y.Q. Polarimetric SAR Image Classification Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1935–1939. [Google Scholar] [CrossRef]
  22. Feng, W.; Sui, H.; Huang, W.; Xu, C.; An, K. Water Body Extraction From Very High-Resolution Remote Sensing Imagery Using Deep U-Net and a Superpixel-Based Conditional Random Field Model. IEEE Geosci. Remote Sens. Lett. 2019, 16, 618–622. [Google Scholar] [CrossRef]
  23. Yuan, K.; Zhuang, X.; Schaefer, G.; Feng, J.; Guan, L.; Fang, H. Deep-Learning-Based Multispectral Satellite Image Segmentation for Water Body Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7422–7434. [Google Scholar] [CrossRef]
  24. Chan, L.; Hosseini, M.S.; Plataniotis, K.N. A Comprehensive Analysis of Weakly-Supervised Semantic Segmentation in Different Image Domains. Int. J. Comput. Vis. 2021, 129, 361–384. [Google Scholar] [CrossRef]
  25. Peng, C.; Li, Y.; Jiao, L.; Chen, Y.; Shang, R. Densely Based Multi-Scale and Multi-Modal Fully Convolutional Networks for High-Resolution Remote-Sensing Image Semantic Segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2612–2626. [Google Scholar] [CrossRef]
  26. Tao, A.; Karan, N.; Nvidia, S.; Catanzaro Nvidia, B. Hierarchical Multi-Scale Attention For Semantic Segmentation. arXiv 2020, arXiv:2005.10821. [Google Scholar] [CrossRef]
  27. Yu, B.; Yang, L.; Chen, F. Semantic Segmentation for High Spatial Resolution Remote Sensing Images Based on Convolution Neural Network and Pyramid Pooling Module. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3252–3261. [Google Scholar] [CrossRef]
  28. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar] [CrossRef] [Green Version]
  29. Kampffmeyer, M.; Salberg, A.B.; Jenssen, R. Semantic Segmentation of Small Objects and Modeling of Uncertainty in Urban Remote Sensing Images Using Deep Convolutional Neural Networks. In Proceedings of the Pattern Recognition Workshops (CVPRW), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 680–688. [Google Scholar] [CrossRef]
  30. Sun, W.; Wang, R. Fully Convolutional Networks for Semantic Segmentation of Very High Resolution Remotely Sensed Images Combined With DSM. IEEE Geosci. Remote Sens. Lett. 2018, 15, 474–478. [Google Scholar] [CrossRef]
  31. Jiang, C.; Zhang, H.; Wang, C.; Ge, J.; Wu, F. Water Surface Mapping from Sentinel-1 Imagery Based on Attention-UNet3+: A Case Study of Poyang Lake Region. Remote Sens. 2022, 14, 4708. [Google Scholar] [CrossRef]
  32. Sarp, S.; Kuzlu, M.; Cetin, M.; Sazara, C.; Guler, O. Detecting Floodwater on Roadways from Image Data Using Mask-R-CNN. In Proceedings of the 2020 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Novi Sad, Serbia, 24–26 August 2020; pp. 1–6. [Google Scholar] [CrossRef]
  33. Zhao, J.; Pelich, R.; Hostache, R.; Margen, P.; Cao, S.; Wagner, W.; Chini, M. Deriving an Exclusion Map (Ex-Map) from Sentinel-l Time Series for Supporting Floodwater Mapping. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 1280–1283. [Google Scholar] [CrossRef]
  34. Sazara, C.; Cetin, M.; Iftekharuddin, K.M. Detecting floodwater on roadways from image data with handcrafted features and deep transfer learning. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 804–809. [Google Scholar] [CrossRef] [Green Version]
  35. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and flexible image augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef] [Green Version]
  36. Maggiori, E.; Tarabalka, Y.; Charpiat, G.; Alliez, P. Can semantic labeling methods generalize to any city? The inria aerial image labeling benchmark. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 3226–3229. [Google Scholar] [CrossRef]
  37. Frolov, V.; Faizov, B.; Shakhuro, V.; Sanzharov, V.; Konushin, A.; Galaktionov, V.; Voloboy, A. Image Synthesis Pipeline for CNN-Based Sensing Systems. Sensors 2022, 22, 2080. [Google Scholar] [CrossRef]
  38. Dumoulin, V.; Visin, F. A guide to convolution arithmetic for deep learning. arXiv 2016, arXiv:1603.07285. [Google Scholar] [CrossRef]
  39. Hamwood, J.; Alonso-Caneiro, D.; Read, S.A.; Vincent, S.J.; Collins, M.J. Effect of patch size and network architecture on a convolutional neural network approach for automatic segmentation of OCT retinal layers. Biomed. Opt. Express 2018, 9, 3049. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar] [CrossRef]
  41. Elgeldawi, E.; Sayed, A.; Galal, A.R.; Zaki, A.M. Hyperparameter tuning for machine learning algorithms used for arabic sentiment analysis. Informatics 2021, 8, 79. [Google Scholar] [CrossRef]
  42. Jin, H.; Song, Q.; Hu, X. Auto-Keras: An Efficient Neural Architecture Search System. In Proceedings of the Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; ACM: New York, NY, USA, 2019; pp. 1946–1956. [Google Scholar] [CrossRef] [Green Version]
  43. Zeiler, M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv 2012, arXiv:1212.5701. [Google Scholar]
  44. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. arXiv 2017, arXiv:1708.02002. [Google Scholar]
  45. Goutte, C.; Gaussier, E. A Probabilistic Interpretation of Precision, Recall and F-Score, with Implication for Evaluation. In Proceedings of the Advances in Information Retrieval, Santiago de Compostela, Spain, 21–23 March 2005; Springer International Publishing: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar] [CrossRef]
  46. Nemni, E.; Bullock, J.; Belabbes, S.; Bromley, L. Fully Convolutional Neural Network for Rapid Flood Segmentation in Synthetic Aperture Radar Imagery. Remote Sens. 2020, 12, 2532. [Google Scholar] [CrossRef]
  47. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer International Publishing: Cham, Switzerland, 2017; pp. 240–248. [Google Scholar] [CrossRef] [Green Version]
  48. Jadon, S. A survey of loss functions for semantic segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Virtual, 27–29 October 2020; pp. 1–7. [Google Scholar] [CrossRef]
  49. Xiao, A.; Yang, X.; Lu, S.; Guan, D.; Huang, J. FPS-Net: A convolutional fusion network for large-scale LiDAR point cloud segmentation. ISPRS J. Photogramm. Remote Sens. 2021, 176, 237–249. [Google Scholar] [CrossRef]
  50. Wang, S.; Liu, W.; Wu, J.; Cao, L.; Meng, Q.; Kennedy, P.J. Training deep neural networks on imbalanced data sets. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 4368–4374. [Google Scholar] [CrossRef]
  51. Li, Z.; Kamnitsas, K.; Glocker, B. Analyzing Overfitting Under Class Imbalance in Neural Networks for Image Segmentation. IEEE Trans. Med. Imaging 2021, 40, 1065–1077. [Google Scholar] [CrossRef]
  52. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  53. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  54. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  55. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for volumetric medical image segmentation. In Proceedings of the 2016 4th International Conference on 3D Vision, 3DV 2016, Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar] [CrossRef] [Green Version]
  56. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R.S., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  57. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  58. Chaurasia, A.; Culurciello, E. LinkNet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar] [CrossRef] [Green Version]
  59. Roh, Y.; Heo, G.; Whang, S.E. A Survey on Data Collection for Machine Learning: A Big Data—AI Integration Perspective. IEEE Trans. Knowl. Data Eng. 2021, 33, 1328–1347. [Google Scholar] [CrossRef]
Figure 1. VV and VH are plotted in grayscale mode, while DEM and Ground Truth (GR) are plotted using Earthply package and matplotlib, respectively. (c) An ideal case with consistent water features and mask (label); (a) shows that features and mask have significant variation; and (b) illustrates that some parts of the mask image do not match features well, especially when comparing with DEM. In the fourth column, yellow indicates water label and purple denotes background. Unassigned pixels are labeled as non-water in the mask, e.g., (c) bottom right corner.
Figure 1. VV and VH are plotted in grayscale mode, while DEM and Ground Truth (GR) are plotted using Earthply package and matplotlib, respectively. (c) An ideal case with consistent water features and mask (label); (a) shows that features and mask have significant variation; and (b) illustrates that some parts of the mask image do not match features well, especially when comparing with DEM. In the fourth column, yellow indicates water label and purple denotes background. Unassigned pixels are labeled as non-water in the mask, e.g., (c) bottom right corner.
Sensors 22 08245 g001
Figure 2. The data is passed to DFM for normalization. The normalized data is processed by the data augmentation library and divided into training, validation and testing samples. In the PHR and PHR-CB experiments, the data is passed through the NAP augmentation module after DFM but for the CFR, CFR-CB experiments, NAP is not required.
Figure 2. The data is passed to DFM for normalization. The normalized data is processed by the data augmentation library and divided into training, validation and testing samples. In the PHR and PHR-CB experiments, the data is passed through the NAP augmentation module after DFM but for the CFR, CFR-CB experiments, NAP is not required.
Sensors 22 08245 g002
Figure 3. The FAPNET model takes the DFM output and passes data through or skips the NAP augmentation module, depending on the experimental setup. Each layer’s downblock output and the previous layer’s upblock are concatenated and the final pixel-wise prediction is provided by CNN with softmax. Between the CNN layers, each down-block and up-block has a dropout layer with a value of 0.2 .
Figure 3. The FAPNET model takes the DFM output and passes data through or skips the NAP augmentation module, depending on the experimental setup. Each layer’s downblock output and the previous layer’s upblock are concatenated and the final pixel-wise prediction is provided by CNN with softmax. Between the CNN layers, each down-block and up-block has a dropout layer with a value of 0.2 .
Sensors 22 08245 g003
Figure 4. Pass Feature Fusion (FF) and Mask to the Training Loop based on the experimental setup, using either DFM or a combination of DFM and NAP augmentation module. The loss is calculated using Focal Loss and the model parameters are updated using Gradient. The model is evaluated using MeanIoU as it is most commonly used by researchers. However, for reference purposes, we also report the Precision, Recall, F1 and Dice Coefficient.
Figure 4. Pass Feature Fusion (FF) and Mask to the Training Loop based on the experimental setup, using either DFM or a combination of DFM and NAP augmentation module. The loss is calculated using Focal Loss and the model parameters are updated using Gradient. The model is evaluated using MeanIoU as it is most commonly used by researchers. However, for reference purposes, we also report the Precision, Recall, F1 and Dice Coefficient.
Sensors 22 08245 g004
Figure 5. Experimental result comparison on the datasets using the MeanIoU score. In the CFR-CB and PHR-CB experiments (both with class balance), our FAPNET gives better MeanIoU scores than the other models and other experiments with class imbalance samples.
Figure 5. Experimental result comparison on the datasets using the MeanIoU score. In the CFR-CB and PHR-CB experiments (both with class balance), our FAPNET gives better MeanIoU scores than the other models and other experiments with class imbalance samples.
Sensors 22 08245 g005
Figure 6. First four rows show VV, VH, DEM and GR, respectively. VV, VH are plotted in grayscale mode, while DEM and GR are plotted using the Earthply package and matplotlib, respectively. A DEM represents ground elevations, where a color indicates the elevation at the pixel location. The color scale ranges from the minimum elevation value (blue) to the maximum elevation value (brown). Green and Yellow represent the elevations in between. We focus on blue for our water-body detection task. The remaining rows indicate different model predictions and the last row is our proposed FAPNET model prediction from the test dataset. Three randomly selected examples are presented for comparison between existing models and our proposed FAPNET model. Each column under an example represents an experimental setup described in Section 4.2, Section 4.3, Section 4.4 and Section 4.5, respectively.
Figure 6. First four rows show VV, VH, DEM and GR, respectively. VV, VH are plotted in grayscale mode, while DEM and GR are plotted using the Earthply package and matplotlib, respectively. A DEM represents ground elevations, where a color indicates the elevation at the pixel location. The color scale ranges from the minimum elevation value (blue) to the maximum elevation value (brown). Green and Yellow represent the elevations in between. We focus on blue for our water-body detection task. The remaining rows indicate different model predictions and the last row is our proposed FAPNET model prediction from the test dataset. Three randomly selected examples are presented for comparison between existing models and our proposed FAPNET model. Each column under an example represents an experimental setup described in Section 4.2, Section 4.3, Section 4.4 and Section 4.5, respectively.
Sensors 22 08245 g006
Figure 7. First, second and third row indicate the input feature, ground truth and predicted mask by FAPNET, respectively, for the PHR-CB experiment. VV and VH are plotted in grayscale mode, while DEM, GR, DFM and FAPNET predictions are plotted using Earthply package and matplotlib, respectively. A DEM represents ground elevations, where a color indicates the elevation at the pixel location. The color scale ranges from the minimum elevation value (blue) to the maximum elevation value (brown). Green and Yellow represent the elevations in between. We focus on blue for our water-body detection task. Pink/red and dark blue color in the DFM indicate background and water respectively Two examples are selected which we compared earlier, in Figure 6.
Figure 7. First, second and third row indicate the input feature, ground truth and predicted mask by FAPNET, respectively, for the PHR-CB experiment. VV and VH are plotted in grayscale mode, while DEM, GR, DFM and FAPNET predictions are plotted using Earthply package and matplotlib, respectively. A DEM represents ground elevations, where a color indicates the elevation at the pixel location. The color scale ranges from the minimum elevation value (blue) to the maximum elevation value (brown). Green and Yellow represent the elevations in between. We focus on blue for our water-body detection task. Pink/red and dark blue color in the DFM indicate background and water respectively Two examples are selected which we compared earlier, in Figure 6.
Sensors 22 08245 g007
Figure 8. Performance comparison with the competition’s 1st, 2nd and 3rd outcomes.
Figure 8. Performance comparison with the competition’s 1st, 2nd and 3rd outcomes.
Sensors 22 08245 g008
Table 1. Over five years, 542 chips or 1084 images from 13 geographic regions were acquired. Each chip contains VV and VH information.
Table 1. Over five years, 542 chips or 1084 images from 13 geographic regions were acquired. Each chip contains VV and VH information.
LocationNo of ChipMonthYear
Bolivia15February2019
Cambodia30August2018
Ghana53September2018
India68August2016
Nigeria18September2018
Pakistan28June2017
Paraguay66October2018
Slovakia65October2020
Somalia26May2018
Spain30September2019
Sri Lanka42May2017
UK32February2019
USA69May2019
Total542
Table 2. FAPNET model layer construction for the CFR and CFR-CB experiments.
Table 2. FAPNET model layer construction for the CFR and CFR-CB experiments.
LayersConv2D 1 Conv2D 2 PaddingNon-Linearities
D a t a F u s i o n (512, 512, 3)---
d o w n b l o c k -1(512, 512, 16)(512, 512, 16)SameReLU
d o w n b l o c k -2(256, 256, 32)(256, 256, 32)SameReLU
d o w n b l o c k -3(128, 128, 64)(128, 128, 64)SameReLU
d o w n b l o c k -4(64, 64, 128)(64, 64, 128)SameReLU
d o w n b l o c k -5(32, 32, 256)(32, 32, 32)SameReLU
d o w n b l o c k -6(16, 16, 512)(16, 16, 64)SameReLU
m i d d l e b l o c k (8, 8, 1012)(8, 8, 256)SameReLU
u p b l o c k -6(16, 16, 128)(16, 16, 128)SameReLU
u p b l o c k -5(32, 32, 128)(32, 32, 128)SameReLU
u p b l o c k -4(64, 64, 128)(64, 64, 128)SameReLU
u p b l o c k -3(128, 128, 64)(128, 128, 64)SameReLU
u p b l o c k -2(256, 256, 32)(256, 256, 32)SameReLU
u p b l o c k -1(512, 512, 16)(512, 512, 16)SameReLU
C o n v 2 D (512, 512, 2)--Softmax
Table 3. FAPNET model layer construction for the PHR and PHR-CB experiments.
Table 3. FAPNET model layer construction for the PHR and PHR-CB experiments.
LayersConv2D 1 Conv2D 2 PaddingNon-Linearities
D a t a F u s i o n + N a p (256, 256, 3)---
d o w n b l o c k -1(256, 256, 16)(256, 256, 16)SameReLU
d o w n b l o c k -2(128, 128, 32)(128, 128, 32)SameReLU
d o w n b l o c k -3(64, 64, 64)(64, 64, 64)SameReLU
d o w n b l o c k -4(32, 32, 128)(32, 32, 128)SameReLU
d o w n b l o c k -5(16, 16, 256)(16, 16, 32)SameReLU
d o w n b l o c k -6(8, 8, 512)(8, 8, 64)SameReLU
m i d d l e b l o c k (4, 4, 1012)(4, 4, 256)SameReLU
u p b l o c k -6(8, 8, 128)(8, 8, 128)SameReLU
u p b l o c k -5(16, 16, 128)(16, 16, 128)SameReLU
u p b l o c k -4(32, 32, 128)(32, 32, 128)SameReLU
u p b l o c k -3(64, 64, 64)(64, 64, 64)SameReLU
u p b l o c k -2(128, 128, 32)(128, 128, 32)SameReLU
u p b l o c k -1(256, 256, 16)(256, 256, 16)SameReLU
C o n v 2 D (256, 256, 2)--Softmax
Table 4. Pixel-wise distribution across the training, validation and testing dataset, with class weights listed in the last column.
Table 4. Pixel-wise distribution across the training, validation and testing dataset, with class weights listed in the last column.
ClassTrainingValidationTestClass Weights
Water15,804,5632,611,5091,941,9700.14
Background97,703,78911,544,26712,475,9500.86
Total113,508,35214,155,77614,417,9201.0
Table 5. CFR experimental results on floodwater dataset [8]. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset column indicate first, second and third best, respectively.
Table 5. CFR experimental results on floodwater dataset [8]. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset column indicate first, second and third best, respectively.
Focal LossMIoUDice CoffF1PrecisionRecall
ModelTrnValTstTrnValTstTrnValTstTrnValTstTrnValTstTrnValTst
UNET [54]0.07260.07880.07670.90840.84670.83060.98230.94970.95670.95270.87920.84550.96350.88450.89610.94780.87610.8370
VNET [55]0.07790.07820.08030.81580.84510.74630.94810.94680.93530.85080.84780.74020.88270.93850.89950.84600.81470.7119
DNCNN [52]0.07610.07690.07750.84550.86370.81320.95820.95360.94950.84140.80050.79150.87980.84240.89320.83770.80570.7723
UNET++ [56]0.07830.07630.07910.80630.85570.77770.94780.95050.93990.77540.75100.74440.82470.80650.85770.79410.74700.7365
U2NET [57]0.07280.07790.07960.90710.85000.76390.97540.95210.93890.84370.78680.74630.88880.86230.88660.84200.77790.7320
ATTUNET [53]0.07700.07770.07960.82250.85450.77170.95460.95110.93970.78350.75650.73930.84690.80480.87670.79330.75760.7224
FPN [49]0.07380.07480.07690.88450.87430.84500.96990.95580.95780.88800.85380.85410.91490.88730.91130.88030.84500.8341
LINKNET [58]0.07580.07750.07810.84650.85460.80210.96010.95130.94840.79470.77200.77350.86060.82710.89650.79560.77500.7567
FAPNET0.07130.07320.07550.92850.89370.85440.97600.96010.96300.93710.89480.85580.94910.91330.90710.95930.88350.8363
Table 6. CFR-CB experimental results on floodwater dataset. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset column indicate first, second and third best, respectively.
Table 6. CFR-CB experimental results on floodwater dataset. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset column indicate first, second and third best, respectively.
Focal LossMIoUDice CoffF1PrecisionRecall
ModelTrnValTstTrnValTstTrnValTstTrnValTstTrnValTstTrnValTst
UNET [54]0.10090.10510.07600.93100.85630.84560.84570.82620.95900.73900.71680.83710.83190.79100.87630.71050.70510.8357
VNET [55]0.10320.10540.08020.86590.84630.75510.83210.82310.93430.67450.68750.75270.78480.80830.87390.65950.67730.7357
DNCNN [52]0.10360.10430.07740.85670.87020.80500.82910.82260.94600.65260.64950.80100.77100.73530.88590.64560.64770.7840
UNET++ [56]0.10400.10470.07780.83020.85560.81980.82760.82190.94250.60900.61560.78880.73460.71810.85410.61090.62260.7913
U2NET [57]0.10180.10470.07830.90140.85640.78980.84020.82530.94750.63640.61740.76110.77980.72420.88150.63370.63480.7426
ATTUNET [53]0.10250.10140.07990.83920.85070.75790.96010.95150.93700.79860.76490.75920.85640.83590.87980.85640.76500.7497
FPN [49]0.10330.10340.07480.86540.87890.86510.83240.83040.96180.66790.68390.82920.77900.77140.90320.65580.67970.8095
LINKNET [58]0.10050.10660.07580.93320.89590.86380.84690.96540.96500.68970.85980.84910.82040.90080.91180.67300.85960.8382
FAPNET0.10020.10420.07520.92640.87970.86790.85560.84090.96680.72760.72400.83220.84170.81310.92190.69950.71160.8125
Table 7. PHR experimental results on floodwater dataset. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset column indicate first, second and third best, respectively.
Table 7. PHR experimental results on floodwater dataset. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset column indicate first, second and third best, respectively.
Focal LossMIoUDice CoffF1PrecisionRecall
ModelTrnValTstTrnValTstTrnValTstTrnValTstTrnValTstTrnValTst
UNET [54]0.07730.07720.07760.82410.85900.80580.95230.95300.95110.81190.68790.76720.86690.75020.84520.80700.80210.7908
VNET [55]0.07790.08030.08000.81510.82540.75520.95010.94030.94090.79790.67300.68440.85420.73810.83770.79960.79860.7027
DNCNN [52]0.07740.07780.07810.82660.85520.79590.95290.94260.94730.80900.66490.72150.86420.70280.80990.80470.78450.7478
UNET++ [56]0.07750.07690.07840.82110.85970.78780.95180.95390.94590.79770.65630.70480.84880.75630.83450.80100.75990.7276
U2NET [57]0.07510.07650.07810.86710.87190.80450.96520.95840.95130.81820.67510.73500.87980.74790.83880.81370.79250.7666
ATTUNET [53]0.07870.07790.07990.80280.84770.74960.94670.94920.94000.78570.65310.68460.84680.71810.80760.78890.76490.7084
FPN [49]0.07270.07470.07620.90650.89410.83010.97570.96560.95900.87470.68910.76600.91260.76740.87630.86280.78700.7743
LINKNET [58]0.07250.07520.07550.91080.89040.84540.97660.96370.96210.88510.71010.77560.92080.75800.87060.87480.81850.7891
FAPNET0.07730.07720.07960.82410.85900.76490.95230.95300.94320.81190.68790.72160.86690.75020.83490.80700.80210.7498
Table 8. PHR-CB experimental results on floodwater dataset. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset indicate first, second and third best, respectively.
Table 8. PHR-CB experimental results on floodwater dataset. Average result was taken after 150 epochs with a ± 2 . 50 % error rate. Green, Yellow and Blue in the MeanIoU Test dataset indicate first, second and third best, respectively.
Focal LossMIoUDice CoffF1PrecisionRecall
ModelTrnValTstTrnValTstTrnValTstTrnValTstTrnValTstTrnValTst
UNET [54]0.07330.08210.08030.94800.87170.79300.97200.93210.94290.96830.86170.75220.96860.86670.77780.96800.91400.8311
VNET [55]0.08350.08400.08260.85890.85450.75730.91910.91770.92340.90810.79980.72840.90810.83320.77800.91120.86180.8037
DNCNN [52]0.08440.08020.08030.84960.88550.78440.91410.93730.93800.90230.82830.75500.90390.85030.81280.90680.88650.8146
UNET++ [56]0.07690.08020.08200.91530.88670.76640.95360.93890.92860.94600.79190.74340.94640.82750.79240.94640.86820.8093
U2NET [57]0.07480.07920.08300.94810.86390.76150.97410.94490.92360.96630.77660.74230.96670.80930.76810.96680.87920.8353
ATTUNET [53]0.07760.08000.08080.90950.88730.77760.94990.93900.93470.94060.79930.73760.94130.82620.77200.94110.87040.7993
FPN [49]0.07520.08000.09010.93110.88810.69670.96270.93950.89000.95470.80980.70820.95460.82150.71110.95580.88790.8497
LINKNET [58]0.07890.07930.08900.89850.89410.69960.94420.94250.89460.93420.80470.72130.93450.81760.73000.93550.86020.8430
FAPNET0.07300.07920.07940.95140.89600.87060.96950.94540.94000.96120.87060.75800.95650.87950.80640.95840.89670.8508
Table 9. Comparison of Time and Memory Consumption during Training up to epoch 5 without compromising the accuracy. Longer training times and more computational overhead are associated with a higher refactor value. A U2NET refactoring value of 1 indicates that the model’s parameters are larger than those of competing models.
Table 9. Comparison of Time and Memory Consumption during Training up to epoch 5 without compromising the accuracy. Longer training times and more computational overhead are associated with a higher refactor value. A U2NET refactoring value of 1 indicates that the model’s parameters are larger than those of competing models.
ModelRefactorBatchGPU
(gb)
RAM
(gb)
Batch Time
(s)
Time for
5 Epoch (min)
UNET [54] 0.04 10 4.6 3.2 79 6.70
VNET [55] 0.33 10 7.9 3.5 95 8.08
DNCNN [52] 0.01 10 7.9 3.5 201 17.26
UNET++ [56] 0.15 10 7.9 3.4 163 13.77
U2NET [57]16 7.9 4.2 273 23.49
ATTUNET [53] 0.19 10 7.9 3.5 194 16.33
FPN [49] 0.13 10 6.6 3.4 128 10.93
LINKNET [58] 0.12 10 4.6 3.6 83 7.11
FAPNET 0.13 10 7.9 3.3 79 6.77
Table 10. Experimental results for individual input feature on floodwater dataset. Average result was taken after 150 epochs with a ± 2.50 % error rate. Green, Yellow and Blue indicate first, second and third best, respectively, for the test dataset in each row (MeanIoU).
Table 10. Experimental results for individual input feature on floodwater dataset. Average result was taken after 150 epochs with a ± 2.50 % error rate. Green, Yellow and Blue indicate first, second and third best, respectively, for the test dataset in each row (MeanIoU).
MetricsVVVHDEMDFM
TrnValTstTrnValTstTrnValTstTrnValTst
MeanIoULoss0.86000.78000.54050.84000.82000.60630.49000.52000.54040.95140.89600.8706
0.48000.31000.11250.48000.29000.10220.65000.69000.12800.07300.07920794
Table 11. Comparison of Time and Memory Consumption during Training up to epoch 5 without compromising the accuracy for different input features VV, VH, DEM and DFM.
Table 11. Comparison of Time and Memory Consumption during Training up to epoch 5 without compromising the accuracy for different input features VV, VH, DEM and DFM.
ModelBatchGPU
(gb)
RAM
(gb)
Batch Time
(ms)
Time for
5 Epoch (min)
VV10 4.6 3.3 39 3.39
VH10 4.6 3.3 42 3.69
DEM10 4.6 3.3 34 2.94
DFM10 7.9 3.3 79 6.77
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Islam, M.S.; Sun, X.; Wang, Z.; Cheng, I. FAPNET: Feature Fusion with Adaptive Patch for Flood-Water Detection and Monitoring. Sensors 2022, 22, 8245. https://doi.org/10.3390/s22218245

AMA Style

Islam MS, Sun X, Wang Z, Cheng I. FAPNET: Feature Fusion with Adaptive Patch for Flood-Water Detection and Monitoring. Sensors. 2022; 22(21):8245. https://doi.org/10.3390/s22218245

Chicago/Turabian Style

Islam, MD Samiul, Xinyao Sun, Zheng Wang, and Irene Cheng. 2022. "FAPNET: Feature Fusion with Adaptive Patch for Flood-Water Detection and Monitoring" Sensors 22, no. 21: 8245. https://doi.org/10.3390/s22218245

APA Style

Islam, M. S., Sun, X., Wang, Z., & Cheng, I. (2022). FAPNET: Feature Fusion with Adaptive Patch for Flood-Water Detection and Monitoring. Sensors, 22(21), 8245. https://doi.org/10.3390/s22218245

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop