Open access peer-reviewed chapter

Despeckling of Multitemporal Sentinel SAR Images and Its Impact on Agricultural Area Classification

Written By

Vladimir Lukin, Oleksii Rubel, Ruslan Kozhemiakin, Sergey Abramov, Andrii Shelestov, Mykola Lavreniuk, Mykola Meretsky, Benoit Vozel and Kacem Chehdi

Submitted: 02 March 2017 Reviewed: 20 November 2017 Published: 20 December 2017

DOI: 10.5772/intechopen.72577

From the Edited Volume

Recent Advances and Applications in Remote Sensing

Edited by Ming-Chih Hung and Yi-Hwa Wu

Chapter metrics overview

1,601 Chapter Downloads

View Full Metrics

Abstract

This chapter addresses an important practical task of classification of multichannel remote sensing data with application to multitemporal dual-polarization Sentinel radar images acquired for agricultural regions in Ukraine. We first consider characteristics of dual-polarization Sentinel radar images and discuss what kind of filters can be applied to such data. Several examples of denoising are presented with analysis of what properties of filters are desired and what can be provided in practice. It is also demonstrated that the use of preliminary denoising produces improvement of classification accuracy where despeckling that is more efficient in terms of standard filtering criteria results in better classification.

Keywords

  • Sentinel
  • dual polarization
  • radar image
  • classification
  • agricultural area

1. Introduction

Remote sensing (RS) systems have become tools of intensive everyday use for numerous applications in ecology, forestry, oceanology, agriculture, and disaster monitoring [1, 2, 3, 4, 5]. This is due to significant advances and achievements in science and technology. Firstly, imagers’ characteristics such as spatial resolution, swath width, receiver input signal-to-noise ratio, bandwidth and throughput, and near real-time data delivery have improved considerably. Secondly, it has become possible to process more data both onboard and on-land in real time or quickly enough. Thirdly, many existing spaceborne RS systems provide data with frequency needed (convenient) for performing monitoring for scenes under interest.

In particular, such monitoring is possible using RS data by Sentinel-1 A/B two-polarization synthetic aperture radar (SAR) that has been put into operation recently [6]. There are several positive features of data (images) provided by this SAR. Firstly, radars are known to be able to operate well (to acquire images) during day and night in almost all weather conditions [7, 8]. Secondly, polarization radars (Sentinel SAR produces VV and VH polarization images) offer more opportunities for effectively solving different classification tasks than single-channel radars [8, 9]. Thirdly, Sentinel SAR is characterized by spatial resolution about 10 m, and data are available free of charge for noncommercial use. Finally, Sentinel satellites carry out global coverage with 6-day revisit frequency.

All together, these facilities of Sentinel SAR provide excellent prerequisites for solving tasks of agricultural monitoring where multitemporal radar images can be employed either separately or jointly with other types of images including optical and infrared ones [10, 11, 12]. However, radar images also have certain drawbacks. One of them is the presence of noise-like phenomenon called speckle [7, 13]. Although speckle intensity and probability density function (PDF) can be different depending upon the number of looks and other factors [7, 14, 15], speckle is the main factor that deteriorates radar image quality and prevents their efficient processing [16, 17] and classification [18].

Taking this into account, speckle noise has to be reduced. Speckle removal (despeckling) is not an easy task. There are several known peculiarities of speckle. Firstly, it is supposed to be pure multiplicative noise [7, 13]. Secondly, it has PDF that is usually non-Gaussian [7, 16]. Thirdly, speckle often possesses spatial correlation [19] that has to be taken into account.

There are many existing image processing software packages, which allow performing radar image pre-filtering such as ESA SNAP toolbox, ENVI, etc. In particular, they provide such good filters as Frost, Lee, refined Lee, etc. [6, 13]. However, there are three aspects worth taking into consideration. Firstly, selection of a proper filter and its parameters (scanning window size, thresholds, etc.) should be done based on careful analysis of speckle properties for SAR data supposed to be used (Sentinel SAR data in our case). Secondly, there are good despeckling filters that have not been yet implemented in toolboxes including filters based on discrete cosine transform (DCT) and/or that belong to a new family of nonlocal filters [20, 21]. Thirdly, efficiency of filtering should be assessed with respect to a final task as crop or agricultural area classification in our case where type of data, methodology of classifier learning, and other factors have essential impact on final classification [9].

The goal of this chapter is threefold. Analysis of speckle characteristics has to be carried out, and the obtained data can be considered as a prerequisite for choosing a proper technique (or several possible methods) of radar image processing. Performance of different despeckling methods has to be evaluated and compared where emphasis should be done on modern techniques. Besides, our intention is to assess the impact of filtering techniques on crop classification accuracy and provide practical recommendations.

Advertisement

2. Image and noise model and its parameters

In this book chapter, we focus on processing Sentinel SAR images. It is known from the literature that the dominant factor degrading SAR image quality is speckle which is supposed to be a specific type of pure multiplicative noise [7, 14], i.e.:

Ikijn=Ikijtrueμkij,i=1,,I,j=1,,J,k=1,,K,E1

where Ikijn denotes the ijth sample of the kth component for a considered multichannel image, μkij is the ijth value of the multiplicative noise in the kth component image, Ikijtrue is the true value for the kijth pixel, I and J define the data size, and K denotes the number of components (K = 2 for two-polarization data that are mainly considered below). Random variable μkij is supposed to have unity mean and variance σμk2 that, in general, can be different for different component images.

To the best of our knowledge, the assumption on pure multiplicative nature of speckle in Sentinel images has not been tested. Other properties of speckle such as spatial correlation have not been thoroughly analyzed yet as well. Thus, we have performed a preliminary analysis. First, our study has been performed for manually selected image fragments. Examples of such fragments (all of rectangular shape) are shown by frames for vertical-vertical (VV) and vertical-horizontal (VH) polarization components of SAR image fragments presented in Figure 1. Analysis has confirmed the assumption that speckle is pure multiplicative. The estimated value of its relative variance σμk2 is about 0.05 for both components (k = 1 and 2), i.e., we deal with multi-look data. Automatic blind estimation [19] has resulted in almost the same estimates of σμk2 that varied only slightly (no more than 10–15% for 19 analyzed VV and 19 studied VH images) (Figure 2).

Figure 1.

An example of Sentinel SAR images in Kiev region of Ukraine with marked homogeneous regions.

Figure 2.

The sensed region (Figure 1) in Ukraine to south from Kiev marked in map.

Visual inspection of images in Figure 1 also shows that they are similar and different simultaneously. The objects might have different contrasts with respect to background or neighbor objects. Some objects can be present in image with one polarization and be absent in image of other polarizations. For example, there is a rectangular shape high-intensity object in the central part of VV image (Figure 1, left) that is not observed in VH image (Figure 1, right).

Cross-correlation factor values for VV and VH polarizations (under condition that they are jointly registered or superimposed with high accuracy) are about 0.8. Speckle has been found practically uncorrelated between component images. These properties will be taken into account in consideration of filters to be applied.

Another important property of speckle is its spatial correlation. There are different approaches to describe and analyze it. One approach is to obtain 2D autocorrelation functions in homogeneous image areas [7, 22]. Under condition of a priori known σμk2, such areas can be detected by comparing relative variance σkrloc2 (calculated by Eq. 2) to Th×σμk2. Here, it is supposed that the considered area is of rectangular shape defined by indices imin;imax in one direction and jmin;jmax in the other direction; Th is threshold that can be set approximately equal to 1.2; I¯kloc is the local mean (calculated by Eq. 3):

σkrloc2=i=iminimaxj=jminjmaxIkijnI¯kloc2/imaximin+1jmaxjmin+1I¯kloc2,E2
I¯kloc=i=iminimaxj=jminjmaxIkijn/imaximin+1jmaxjmin+1.E3

Below we characterize speckle correlation in another manner, using spectrum in DCT domain. The reasons for this will become clear later, in Section 3. Suppose that we have found N homogeneous image blocks of size 8×8 pixels, i.e., blocks for which

σkn2=i=iminimin+7j=jminjmin+7IkijnI¯kn2/64I¯kn2Th×σμk2,E4

where I¯kn is the local mean for the nth block. 2D DCT is then calculated for each nth block with obtaining Dnlm,l=1,,8;m=1,,8 where lm are indices of spatial frequencies, l = 1 and m = 1 relate to DC (direct current) term proportional to the block mean, and alternating current (AC) DCT components with larger l and m relate to higher spatial frequencies. Power spectrum shape estimate is obtained through Eq. 5. Then, the normalized DCT spectrum is obtained through Eq. 6:

D̂plm=n=1NDm2lm/Dn211/n,l=1,,8;m=1,,8,E5
D̂normlm=63D̂plm/l=18m=18D̂plmD̂p11,l=1,,8;m=1,,8,E6

where, in fact, D̂normlm is not of interest for us since it is not taken into account in denoising (see the details in Section 3). There exist also other methods of estimation [23] including blind methods [24, 25].

The obtained estimates are presented in Figure 3 for both polarization components. Bin heights that reflect spectrum values clearly demonstrate that speckle is spatially correlated (spectrum values are considerably smaller for higher spatial frequencies). These correlation properties are quite similar for VV and VH polarizations as well as for different images analyzed.

Figure 3.

Examples of estimated normalized DCT spectra in 8×8 blocks.

Thus, it is possible to state that speckle in Sentinel SAR images is practically pure multiplicative with σμk20.05,k=1,2. Its distribution is not Gaussian—this has been shown by Gaussianity tests. Speckle is spatially correlated and practically independent for different polarizations. All these properties should be taken into consideration at image processing stage.

Advertisement

3. Filters and their impacts on images

First of all, let us recall requirements to filters applied to radar image processing. The main requirement is, certainly, effective suppression of speckle noise. However, edge/detail/texture preservation is important as well. Additional requirement is to retain mean level in homogeneous image regions (some filters do not perform this automatically, without specific correction [26]). This is important since mean in homogeneous regions is strictly connected with calibration [7]. Finally, computational efficiency is important, especially in the considered application since standard size of Sentinel SAR images we deal with is large, ~25k × 16k pixels. Note that some requirements are contradictory [27]. In particular, it is difficult to provide a trade-off between speckle suppression and edge/detail/texture preservation.

If we deal with multichannel images (multispectral, hyperspectral, multipolarization ones), there are two main approaches to their filtering [28]. The first is component-wise processing, i.e., if each component image is processed as single-channel image [13]. This approach is, in general, simpler. There exist more methods. It is easier to take into account speckle properties including its spatial spectrum. Processing of multichannel images can be done in parallel.

The second approach is 3D (vectorial) processing [20, 29]. This group of filters is able to take correlation of image components into account. This, under certain conditions, allows improving the efficiency of noise suppression. However, these methods can run into problems of different properties of noise (noise variance can be not the same, spectrum can be not identical, etc.). Then, either special methods of component image preprocessing are needed [30], or the use of 3D processing benefits in full extent becomes impossible.

Below we consider representatives of both groups of methods. All studied filters are based on DCT. The reasons for this are the following. Firstly, DCT is one of the best data decorrelating transforms approaching Karhunen-Loeve transform. This property is important in denoising and data representation. Secondly, DCT denoising is carried out in sliding blocks [31]. Due to this, it is easily adapted to signal-dependent noise [32] and its spatial correlation [28]. Thirdly, 2D DCT is a standard operation in image processing (in particular, in compression), and its efficient software and hardware implementations exist.

Thus, in addition to known despeckling techniques available in ESA SNAP toolbox, we also study filtering techniques based on DCT. For all modifications considered below, the main principles of processing are the same. Denoising is carried out in blocks of square shape, usually of size 8×8 pixels although other variants are possible. In each block, direct 2D DCT is performed. Then, amplitudes of obtained DCT coefficients are compared to thresholds, and one or another type of threshold operation is accomplished (the coefficient with indices (1,1) is remained unchanged in any case). Then, inverse 2D DCT is done, and as the result, filtered values are obtained for all pixels that belong to a given block.

There are different modifications of DCT-based denoising that concern block overlapping and threshold type. Denoising can be done with nonoverlapping, partly overlapping, and fully overlapping blocks. In the latter case, neighbor blocks are shifted by only one pixel with respect to each other in horizontal or vertical direction. Processing with full overlapping of blocks takes more time compared to processing with partly overlapping or nonoverlapping blocks. But the use of fully overlapping blocks provides the best efficiency of filtering in terms of standard quantitative criteria as peak signal-to-noise ratio (PSNR) and visual quality metrics. So, we will further employ this variant of DCT-based denoising. Note that for most image pixels, one has 64 filtered values coming from different block positions. They are averaged (although other methods of aggregation are possible).

There are hard, soft, and combined types of threshold. Below we employ the former type since hard thresholding is the most simple and efficient. This means that if a DCT coefficient is larger than a corresponding threshold, then it is remained unchanged. Otherwise, a zero value is assigned.

Multiplicative nature of speckle and its spatial correlation can be taken into consideration in different manners. Consider possible variants more in detail. If processing is component-wise, signal-dependent (locally adaptive) and frequency-dependent thresholds can be used. The local thresholds in an nth block are determined as

Tnlm=βσnDnorm0.5lm,l=1,,8;m=1,,8,E7

where σn is local noise standard deviation for the nth block that can be approximately calculated as σn=σμI¯n (index k is omitted for simplicity) and β is the filter parameter usually set equal to 2.7 for hard thresholding [32]. This approach to filtering will be further referred as Filter 1.

Another way of component-wise processing mentioned earlier is based on utilizing homomorphic or variance-stabilizing transforms. A general form of logarithmic-type homomorphic transform (HT) is Iijh=clogdIijn [7, 33] where c and d are transform parameters that are both larger than unity. These parameters can be chosen in different ways depending upon certain conditions. The main aspect here is that in multilook SAR images subject to such a HT, speckle converts to pure additive noise with non-Gaussian PDF. Its variance can be calculated as σh2c2σμ2/lnd2.

After this direct HT, a DCT-based filter version for additive noise is applied. Since additive noise is spatially correlated as well and its normalized DCT spectrum is practically the same as for original data, frequency-dependent thresholds are determined as

Tnlm=βσhDnorm0.5lm,l=1,,8;m=1,,8,E8

where the recommended β is the same as earlier. After filtering, inverse HT is performed, and it is desired to carry out mean level correction [26]. This variant of filtering will be further referred as Filter 2. According to our studies, it performs slightly worse compared to Filter 1. One reason is that additional distortions are introduced by direct and inverse HTs.

Joint processing of two component images that we plan to perform for the considered application can be carried out in two ways. The first one [20] is to use one component as basic one and to perform data correction for the second component image with providing equal mean values for each block of these images. Then, 3D DCT is performed in each block, and σn is set according to (4) where I¯n from the basic image is used for its calculation.

Another, practically equivalent, variant is to carry out variance-stabilizing transforms separately for each component images in such a manner that σh12=σh22. If σμ12σμ22 as we have in our case, parameters of logarithmic transform c and d can be the same. After this, 3D is done in each block, then thresholding follows, after this inverse 3D DCT is carried out. Due to separability of 3D transform into two-element 1D DCT and 2D DCT, obtaining of sum and difference images is possible. Then, their separate processing is applied with further inverse two-element DCT for each pixel. Threshold (8) is used in this case. The described denoising method is further called Filter 3.

There is one condition that should be satisfied for both presented 3D filtering techniques. Normalized spectra for component images should be the same or, at least, almost identical. Preliminary analysis carried out in Section 2 has demonstrated that this condition is satisfied. So, we have averaged estimates of normalized spectra obtained for all analyzed fragments of 19 VV and 19 VH radar images.

It is worth recalling that methods for predicting noise suppression efficiency have been proposed previously [27, 34]. Equivalent noise variance can be determined for original image with speckle as

σeqk2=i=1Ij=1Jσμk2Ikijtrue/IJi=1Ij=1Jσμk2Ikijn2/IJ.E9

Then, the proposed methods of efficiency prediction are able to either predict the ratio MSEkout/σeqk2 where

MSEkout=i=1Ij=1JIkijfIkijtrue2/IJ,k=1,,KE10

and Ikijf,i=1,,I;j=1,,J;k=1,,K is the kth component image after despeckling or improvement of PSNR (IPSNR) equal to 10log10σeqk2/MSEkout [35, 36]. Experiments carried out with 512×512 pixel fragments of Sentinel SAR images have shown that IPSNR for Filter 1 varies within the limits from 5 to 12 dB, i.e., images enhance sufficiently.

Let us prove this by two examples. Figure 4a presents original image with VV polarization. The output of Filter 1 is shown in Figure 4b. In turn, Figure 4c represents the output of Filter 2. Finally, Figure 4d shows the output of the modification [30] applied within the scheme [26] instead of standard BM3D (block matching three-dimensional) filter. Note that this is only an example for the latter filter applied to image fragment of limited size. We have not applied it to processing large-size Sentinel data further subject to classification since this filter requires too much time for image denoising.

Figure 4.

Original SAR image (a) and outputs of three denoisers: Filters 1 (b), Filter 2 (c), and nonlocal filter with HT (d).

The filtering results represented in Figure 4 look quite similarly. Efficient speckle suppression is provided, while edge/detail preservation is good enough. Visual inspection carried by us and quantitative data in [30] show that the output of Filter 2 is the most smeared and some artifacts are present. Due to linearity nature of the main used computational procedures (DCT) of the considered filters, all computational costs can be easily reduced using parallel-wised algorithms or hardware. On the following computer configuration—Intel Core i7, 16 GB RAM—the entire calculation time for one image (recall that average size of images is ~25k × 16k pixels) is about 17 min. We also note that such time can be reduced by realizing the considered denoising methods in low-level programming language due to filter simplicity.

We have also analyzed filter outputs for 3D despeckling methods. One example is given in Figure 5. Original image of VV polarization is presented in Figure 5a. Again, a good compromise between speckle suppression and edge/detail preservation is provided.

Figure 5.

Original (noisy) image (a) and outputs of Filter 3 (b) and filters in [20].

However, visual analysis is not able to draw conclusions what filter will suit the final goal of reaching high classification accuracy in the best way. Thus, three denoising techniques (Filters 1, 2, and 3) described above have been used to improve SAR image quality before classification. Note that for multitemporal SAR data, more sophisticated 3D filtering methods can be applied, but our task here is to establish connection between filtering efficiency and classification accuracy. We expect that the use of Filter 2 and Filter 3 will produce comparable or better classification than Filter 1 and other filters available in known toolboxes.

Advertisement

4. Classification of multichannel radar images

There are many classifiers that can be applied. Popular ones such as support vector machine (SVM), decision tree (DT), and random forest (RF) classifiers were often employed for remote sensing image classification in the recent years. Many papers have demonstrated comparable or better performance of neural networks (NN) [37, 38, 39, 40]. Although the NN training can be resource- and time-consuming, the corresponding classifiers have several positive features compared to SVM and DT. In particular, NNs are usually fast at processing new data that is important in processing of large volumes of radar data. Besides, they are able to produce probabilistic outputs, which can be employed for characterizing reliability (quality) of the final results (classification map). Hence, for classification, we have used a specific type of NN—an ensemble of feedforward neural networks, namely, multilayer perceptrons (MLPs). This type of classifiers was validated earlier for five Joint Experiment of Crop Assessment and Monitoring (JECAM) test sites in several countries, and it outperformed other considered techniques [41, 42]. The architecture of the used approach is represented in Figure 6. A committee (ensemble) of MLPs was used to improve performance of individual classifiers [9, 10]. The committee was obtained using a variant of the bagging technique [43] where MLPs with different parameters were trained using the same training data. This approach has several advantages. It is rather simple, non-computationally intensive and proved to be efficient for various applications. Outputs from the used MLPs were integrated using the method of average committee. According to this technique, the average class probability over all elementary classifiers is calculated, and the class that has the highest average posterior probability for the given input sample is chosen [9]. The following equation describes this procedure:

cre=1Qq=1Qprq,c=argmaxc=1,Cpce,E11

where c is the class to which the input sample is assigned by the committee of classifiers, pre denotes the resulting posterior probability of the committee, pce is posterior probability for each of MLPs, Q is the total number of classifiers in the committee, and C denotes the number of classes. There are two differences between average committee procedure and a majority voting technique: (i) the former approach gives probabilistic output, which can be employed as an indicator of reliability for classifying a particular pixel or area; (ii) there is no ambiguity when two or more classes have got the same number of “votes.”

Figure 6.

Architecture of an ensemble of feed forward NNs based on multilayer perceptrons (MLPs).

Let us give some details. Our MLP classifier used hyperbolic tangent activation function for neurons in the hidden layer and logistic activation function in the output layer. We utilized a cross-entropy (CE) error function for training our neural network. This provided better performance in terms of speed of training and classification accuracy [9, 44]:

Ew=lnpTw=s=1Sc=1Ctsclnysc,E12

where E(w) denotes the CE error function that is dependent on the neurons’ weight coefficients w, T is the set of vectors of target outputs in the training set that is composed of S samples, and tsc and ysc denote the target and MLP outputs, respectively. In the target output for class c, all components of vector ts are set to 0, except for the cth component, which is set to 1. The CE error E(w) had to be minimized. This was done by the scaled conjugate gradient algorithm by varying weight coefficients w. Backpropagation method was utilized for NN training. To prevent NN overfitting, we used early stopping and weight decay (L2 regularization) techniques.

Classification of multitemporal radar images was performed on a per-pixel basis [45]. We have considered all SAR images available during the crop growth period. Note that the use of multitemporal images allows more accurate classification of crops than a single-date image.

Advertisement

5. Classification results

Filtering approaches for Sentinel-1 images were verified in the south part of Ukraine (yellow rectangle in Figure 2). Zoomed results after different filters are shown for fragment (marked by star) in Figure 2 that has coordinates (33.73052, 47.18544). Such a territory is an intensive agricultural area; it corresponds to different climatic zones (humid continental) and has many different land cover types (forest on the north and agricultural crop in the southern part). The crop calendar is September–July for winter crops and April–October for spring and summer crops. A typical field size is 30–250 ha.

We preprocessed time series of 10 Sentinel-1A images with different filters available in ESA SNAP toolbox and Filter 1, Filter 2, and Filter 3. The comparison of those filters with non-filtered image for July 7, 2016, is shown in Figure 7. It is not easy to evaluate quality of filtering visually (user’s opinions can be subjective). Therefore, we provide crop classification maps based on the same in situ data for time series of images obtained by each filter. For each time series of images, we independently trained an ensemble of neural networks. For this, we collected 153 data samples for nine classes, and all the accuracies were evaluated on independent set, which consisted of 146 samples for nine classes. Under the term “sample,” we assume one polygon that consists of homogeneous pixels and all of them relate to the same class.

Figure 7.

Example of using different filters applied to Sentinel-1A images (R channel, VV polarization; G channel, VH polarization): (A) image without filtering, (B) refined Lee filter, (C) Filter 1, and (D) Filter 3.

For ground-truth data collecting, we have used the along the roads approach. During these trips, we have collected georeferenced type of crops for definite fields. After that, collected data were divided into two independent datasets: one was used for training and another one for validation (as ground-truth data). So, for training and validation, we have used independent datasets with 153 and 146 samples, respectively.

The obtained crop classification maps after applying different filters are shown in Figure 8. In Tables 1 and 2, we present the comparison of user accuracy (UA), producer accuracy (PA), overall accuracy (OA), and Kappa coefficient for all classes [46] from crop classification map using different filters for SAR data in 2016.

Figure 8.

Example of crop classification maps based on different filters for Sentinel-1A images: (A) without filtering, (B) refined Lee filter, (C) Filter 1, and (D) Filter 3.

ClassNo filterBoxcarFrostGamma mapIDANLee
UA, %PA, %UA, %PA, %UA, %PA, %UA, %PA, %UA, %PA, %UA, %PA, %
Artificial4.642.110.843.115.742.17.143.15.95610.243.5
Winter wheat65.573.579.776.378.476.979.67671.967.379.576.6
Maize62.749.879.654.578.155.478.755.893.262.579.856.2
Sunflower96.590.296.692.996.592.996.792.496.594.796.792.9
Soybeans3761.341.965.642.665.139.665.251.569.241.365.4
Forest52.988.355.492.156.492.961.591.923.94254.992.5
Grassland38.583.644.392.144.99244.385.54893.347.192.1
Bare land49.823.2685167.452.471.346.636.922.460.847.4
Water98.199.997.610098.799.993.210099.510099.1100
OA, %/kappa82.6/0.786.4/0.7686.7/0.7786/0.7586.2/0.7686.5/0.76

Table 1.

Comparison of user accuracy (UA), producer accuracy (PA), overall accuracy (OA), and kappa coefficient (in the lowest raw) for different filters for SAR data in 2016.

ClassLee sigmaMedianRefined LeeDCTFDCTF HT3D DCTF
UA, %PA, %UA, %PA, %UA, %PA, %UA, %PA, %UA, %PA, %UA, %PA, %
Artificial9.443.57.838.96.845.43.752.38.981.98.465.3
Winter wheat69.571.875.87680.377.678.176.185.375.186.170.5
Maize92.661.378.55582.157.992.773.59578.794.982.6
Sunflower96.794.996.592.396.793.897.295.196.895.396.695.2
Soybeans51.669.940.563.745.566.858.171.156.672.559.572.3
Forest45.579.255.393.366.591.948.258.79.519.929.148.4
Grassland53.989.645.891.346.392.844.694.248.890.948.691.1
Bare land36.926.663.54462.647.932.114.923.49.348.629.8
Water95.31009610099.310099.710099.810099.7100
OA, %/kappa87.0/0.7785.8/0.7587.4/0.7887.7/0.7887.5/0.7888.7/0.80

Table 2.

Comparison of user accuracy (UA), producer accuracy (PA), overall accuracy (OA), and kappa coefficient (in the lowest raw) for different filters for SAR data in 2016.

The overall accuracy of the crop classification map without applying any filtering is 82.6%. The lowest accuracy among all the available filters is provided by Median filter; classification accuracy is higher by +3.2% than classification of original (non-filtered) data. The proposed Filter 1 and Filter 2 have similar results with a nonsignificant gain of +0.3 and +0.1% compared to refined Lee, respectively. The most accurate crop map was obtained based on images preprocessed by the Filter 3. The gain of using this method is +6.1% compared to classification of original data and +1.3% compared to the refined Lee filter. It is worth noting that Filter 3 positive gains were observed not only for overall accuracy but also for PA and UA for each class, excluding forest and bare land.

Advertisement

6. Conclusions

One aim of this study was to analyze properties of Sentinel SAR images. It has been shown that speckle is quite intensive and spatially correlated. These peculiarities have been taken into account in choosing proper filters for data preprocessing.

Another aim was to compare the impact of different filters for SAR imagery denoising on crop classification accuracy. Ten SAR images with VV and VH polarization acquired by Sentinel-1A satellite for the Ukraine territory were explored. For speckle suppression, all available filters from the ESA SNAP toolbox were evaluated along with those based on DCT, namely, DCTF, DCTF HT, and 3D DCTF. Filter quality estimation has been performed in terms of overall crop classification accuracy. An ensemble of feed forward neural networks, in particular MLPs, was used as this method was validated for multiple JECAM test sites and proved to be efficient for multitemporal image classification.

The best performance using available filters in SNAP toolbox in terms of overall accuracy was achieved by the refined Lee filter. At the same time, all proposed DCT-based filters outperformed common filters from the SNAP toolbox. 3D DCTF provided the most accurate crop classification map after removing the speckle, and that is essential for crop area estimation and for solving other applied problems. It should be noted that 3D DCTF improved not only overall accuracy but also PA and UA for each class (especially for summer crops), excluding forest and bare land.

References

  1. 1. Schowengerdt R. Remote Sensing: Models and Methods for Image Processing. 3rd ed. Cambridge: Academic Press; 2006. 560 p. DOI: 10.978.00804/80589
  2. 2. Blanes I, Magli E, Serra-Sagrista JA. Tutorial on image compression for optical space imaging systems. IEEE Geoscience and Remote Sensing Magazine. 2014;2(3):8-26. DOI: 10.1109/MGRS.2014.2352465
  3. 3. Kravchenko A, Kussul N, Lupian E, Savorsky V, Hluchy L, Shelestov A. Water resource quality monitoring using heterogeneous data and high-performance computations. Cybernetics and Systems Analysis. 2008;44(4):616-624. DOI: 10.1007/s10559-008-9032-x
  4. 4. Kogan F, Kussul N, Adamenko T, Skakun S, Kravchenko A, Krivobok A, Shelestov A, Kolotii A, Kussul O, Lavrenyuk A. Winter wheat yield forecasting: A comparative analysis of results of regression and biophysical models. Journal of Automation and Information Sciences. 2013;45(6):68-81. DOI: 10.1615/JAutomatInfScien.v45.i6.70
  5. 5. Skakun S, Kussul N, Shelestov A, Kussul O. The use of satellite data for agriculture drought risk quantification in Ukraine. Geomatics, Natural Hazards and Risk. 2015;7(3):901-917. DOI: 10.1080/19475705.2015.1016555
  6. 6. European Space Agency. Earth Online [Internet]. 2006. Available from: https://earth.esa.int/documents/653194/656796/Speckle_Filtering.pdf [Accessed: 2017-06-07]
  7. 7. Oliver C, Quegan S. Understanding Synthetic Aperture Radar Images. USA: SciTech Publishing; 2004. 486 p. DOI: 10.978.18911/21319
  8. 8. Lee JS, Pottier E. Polarimetric Radar Imaging: From Basics to Applications. USA: CRC Press; 2009. 438 p. DOI: 10.1080/01431161.2010.535193
  9. 9. Skakun S, Kussul N, Shelestov A, Lavreniuk M, Kussul O. Efficiency assessment of multitemporal C-band Radarsat-2 intensity and Landsat-8 surface reflectance satellite imagery for crop classification in Ukraine. IEEE Journal of Selected Topics in Applied Earth Observation and Remote Sensing. 2016;9(8):3712-3719. DOI: 10.1109/JSTARS.2015.2454297
  10. 10. Kussul N, Skakun S, Shelestov A, Lavreniuk M, Yailymov B, Kussul O. Regional scale crop mapping using multi-temporal satellite imagery. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2015;XL-7/W3:45-52. DOI: 10.5194/isprsarchives-XL-7-W3-45-2015
  11. 11. Kolotii A, Kussul N, Shelestov A, Skakun S, Yailymov B, Basarab R, Lavreniuk M, Oliinyk T, Ostapenko V. Comparison of biophysical and satellite predictors for wheat yield forecasting in Ukraine. International Archives of the. 2015;XL-7/W3:39-44. DOI: 10.5194/isprsarchives-XL-7-W3-39-2015
  12. 12. Kussul N, Lemoine G, Gallego F, Skakun S, Lavreniuk M, Shelestov A. Parcel-based crop classification in Ukraine using Landsat-8 data and sentinel-1A data. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing. 2016;9(6):2500-2508. DOI: 10.1109/JSTARS.2016.2560141
  13. 13. Touzi R. Review of speckle filtering in the context of estimation theory. IEEE Transactions on Geoscience and Remote Sensing. 2002;40(11):2392-2404. DOI: 10.1109/TGRS.2002.803727
  14. 14. Anfinsen S, Doulgeris A, Eltoft T. Estimation of the equivalent number of looks in polarimetric synthetic aperture radar imagery. IEEE Transactions on Geoscience and Remote Sensing. 2009;47(11):3795-3809. DOI: 10.1109/TGRS.2009.2019269
  15. 15. Lukin V, Ponomarenko N, Abramov S, Vozel B, Chehdi K, Astola J. Filtering of radar images based on blind evaluation of noise characteristics. In: Proceedings of Image and Signal Processing for Remote Sensing XIV (SPIE RS’08); 10 October 2008; Cardiff. SPIE; 2008. p. 12. DOI: 10.1117/12.799396
  16. 16. Kang X, Han C, Yang Y, Tao T. SAR image edge detection by ratio-based Harris method. In: Proceedings of the International Conference of Acoustics, Speech and Signal Processing (ICASSP’06); 14-19 May 2006; Toulouse. IEEE; 2006. pp. 837-840. DOI: 10.1109/ICASSP.2006.1660473
  17. 17. Marques R, Medeiros F, Ushizima D. Target detection in SAR images based on a level set approach. IEEE Transactions on Systems, Man and Cybernetics. 2009;39(2):214-222. DOI: 10.1109/TSMCC.2008.2006685
  18. 18. Lukin V, Ponomarenko N, Fevralev D, Vozel B, Chehdi K, Kurekin A. Classification of pre-filtered multichannel remote sensing images. In: Escalante-Ramirez B, editor. Remote Sensing—Advanced Techniques and Platforms. Austria: InTech; 2012. pp. 75-98. DOI: 10.5772/36046.ch4
  19. 19. Abramov S, Abramova V, Lukin V, Ponomarenko N, Vozel B, Chehdi K, Egiazarian K, Astola J. Methods for blind estimation of speckle variance in SAR images: Simulation results and verification for real-life data. In: Awrejcewicz J, editor. Computational and Numerical Simulations. Austria: InTech; 2014. pp. 303-327. DOI: 10.5772/57035.ch14
  20. 20. Kozhemiakin R, Lukin V, Vozel B, Chehdi K. Filtering of dual-polarization radar images based on discrete cosine transform. In: Proceedings of 15th international radar symposium (IRS’14); 16-18 june 2014; gdansk. IEEE. 2014:1-4. DOI: 10.1109/IRS.2014.6869301
  21. 21. Deledalle CA, Denis L, Poggi G, Tupin F, Verdoliva L. Exploiting patch similarity for SAR image processing: The nonlocal paradigm. IEEE Signal Processing Magazine, Recent Advances in Synthetic Aperture Radar Imaging. 2014;31(4):69-78. DOI: 10.1109/MSP.2014.2311305
  22. 22. Lukin V, Abramov S, Ponomarenko N, Uss M, Zriakhov M, Vozel B, Chehdi K, Astola J. Methods and automatic procedures for processing images based on blind evaluation of noise type and characteristics. SPIE Journal on Advances in Remote Sensing. 2011;5(1):1-27. DOI: 10.1117/1.3539768
  23. 23. Ponomarenko N, Lukin V, Egiazarian K, Astola J. A Method for blind estimation of spatially correlated noise characteristics. In: Proceedings of SPIE Conference Image Processing: Algorithms and Systems VII (SPIE EI’10); 8 February 2010; San Jose. SPIE; 2010. p. 12. DOI: 10.1117/12.847986
  24. 24. Colom M, Buades A, Morel JM. Nonparametric noise estimation method for raw images. Journal of the Optical Society of America. 2014;31(4):863-871. DOI: 10.1364/JOSAA.31.000863
  25. 25. Uss M, Vozel B, Lukin V, Chehdi K. Maximum likelihood estimation of spatially correlated signal-dependent noise in hyperspectral images. Optical Engineering. 2012;51(11):1-12. DOI: 10.1117/1.OE.51.11.111712
  26. 26. Makitalo M, Foi A, Fevralev D, Lukin V. Denoising of single-look SAR images based on variance stabilization and non-local filters. In: Proceedings of International Conference of Mathematical Methods in Electromagnetic Theory (MMET’10); 6-8 September 2010; Kyiv. IEEE; 2010. p. 4. DOI: 10.1109/MMET.2010.5611418
  27. 27. Lukin V, Melnik V, Pogrebniak A, Zelensky A, Saarinen K, Astola J. Digital adaptive robust algorithms for radar image filtering. Journal of Electronic Imaging. 1996;5(3):410-421. DOI: 10.1117/12.240715
  28. 28. Fevralev D, Lukin V, Ponomarenko N, Abramov S, Egiazarian K, Astola J. Efficiency analysis of color image filtering. EURASIP Journal on Advances in Signal Processing. 2011;2011:41. DOI: 10.1186/1687-6180-2011-41
  29. 29. Deledalle CA, Denis L, Tabti S, Tupin F. MuLoG, or how to apply Gaussian denoisers to multi-channel SAR speckle reduction? IEEE Transactions on Image Processing. Forthcoming. DOI: 10.1109/TIP.2017.2713946
  30. 30. Rubel A, Lukin V, Egiazarian K. Block matching and 3D collaborative filtering adapted to additive spatially correlated noise. In: Proceedings of 9th International Workshop on Video Processing and Quality Metrics (VPQM’2015); 26 February 2015; Chandler. 2015. p. 6
  31. 31. Pogrebnyak O, Lukin V. Wiener DCT based image filtering. Journal of Electronic Imaging. 2012;4:1-14. DOI: 10.1117/1.JEI.21.4.043020
  32. 32. Oktem R, Egiazarian K, Lukin V, Ponomarenko N, Tsymbal O. Locally adaptive DCT filtering for signal-dependent noise removal. EURASIP Journal on Advances in Signal Processing. 2007;2007:1-10. DOI: 10.1155/2007/42472
  33. 33. Tsymbal O, Lukin V, Ponomarenko N, Zelensky A, Egiazarian K, Astola J. Three-state locally adaptive texture preserving filter for radar and optical image processing. EURASIP Journal on Applied Signal Processing. 2005;(8):1185-1204. DOI: 10.1155/ASP.2005.1185
  34. 34. Abramov S, Krivenko S, Roenko A, Lukin V, Djurovic I, Chobanu M. Prediction of filtering efficiency for DCT-based image denoising. In: Proceedings of 2nd Mediterranean Conference on Embedded Computing (MECO’2013); 15-20 June 2013; Budva. IEEE; 2013. pp. 97-100. DOI: 10.1109/MECO.2013.6601327
  35. 35. Rubel O, Kozhemiakin R, Abramov S, Lukin V, Vozel B, Chehdi K. Performance prediction for 3D filtering of multichannel images. In: Proceedings of SPIE: Image and Signal Processing for Remote Sensing XXI (SPIE RS’15); 15 October 2015; Toulouse. SPIE; 2015. p. 11. DOI: 10.1117/12.2193976
  36. 36. Rubel O, Lukin V, de Medeiros FS. Prediction of despeckling efficiency of DCT-based filters applied to SAR images. In: Proceedings of International Conference on Distributed Computing in Sensor Systems (DCOSS’15); 10-12 June 2015; Fortaleza. IEEE; 2015. pp. 159-168. DOI: 10.1109/DCOSS.2015.16
  37. 37. Huang FJ, LeCun Y. Large-scale learning with SVM and convolutional NETWORK for generic object recognition. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06); 17-22 June 2006; New York. IEEE; 2006. pp. 284-291. DOI: 10.1109/CVPR.2006.164
  38. 38. Ding J, Chen B, Liu H, Huang M. Convolutional neural network with data augmentation for SAR target recognition. IEEE Geoscience and Remote Sensing Letters. 2016;13(3):364-368. DOI: 10.1109/LGRS.2015.2513754
  39. 39. Kussul N, Lavreniuk M, Skakun S, Shelestov A. Deep learning classification of land cover and crop types using remote sensing data. IEEE Geoscience and Remote Sensing Letters. 2017;14(5):778-782. DOI: 10.1109/LGRS.2017.2681128
  40. 40. Shelestov A, Lavreniuk M, Kussul N, Novikov A, Skakun S. Exploring Google earth engine platform for big data processing: Classification of multi-temporal satellite imagery for crop mapping. Frontiers in Earth Science. 2017;5(17):1-10. DOI: 10.3389/feart.2017.00017
  41. 41. Waldner F, De Abelleyra D, Veron S, Zhang M, Wu B, Plotnikov D, Bartalev S, Lavrenyuk M, Skakun S, Kussul N, Le Marie G, Dupuy S, Jarvis I, Defourny P. Towards a set of agrosystem-specific cropland mapping methods to address the global cropland diversity. International Journal of Remote Sensing. 2016;37(14):3196-3231. DOI: 10.1080/01431161.2016.1194545
  42. 42. Lavreniuk M, Kussul N, Skakun S, Shelestov A, Yailymov B. Regional retrospective high resolution land cover for Ukraine: Methodology and results. In: Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS’15); 26-31 July 2015; Milan. IEEE; 2015. pp. 3965-3968. DOI: 10.1109/IGARSS.2015.7326693
  43. 43. Bishop C. Pattern Recognition and Machine Learning. New York, USA: Springer; 2006. 758 p. DOI: 10.978.03873/10732
  44. 44. Lavreniuk M, Skakun S, Shelestov A, Yalimov B, Yanchevskii S, Yaschuk D, Kosteckiy A. Large-scale classification of land cover using retrospective satellite data. Cybernetics and Systems Analysis. 2016;52(1):127-138. DOI: 10.1007/s10559-016-9807-4
  45. 45. Kussul N, Skakun S, Shelestov A, Kussul O. The use of satellite SAR imagery to crop classification in Ukraine within JECAM project. In: Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS’14); 13-18 July 2014; Quebec City. IEEE; 2014. pp. 1497-1500. DOI: 10.1109/IGARSS.2014.6946721
  46. 46. Congalton R. A review of assessing the accuracy of classifications of remotely sensed data. Remote Sensing of Environment. 1991;37(1):35-46. DOI: 10.1016/0034-4257(91)90048-B

Written By

Vladimir Lukin, Oleksii Rubel, Ruslan Kozhemiakin, Sergey Abramov, Andrii Shelestov, Mykola Lavreniuk, Mykola Meretsky, Benoit Vozel and Kacem Chehdi

Submitted: 02 March 2017 Reviewed: 20 November 2017 Published: 20 December 2017