Open access peer-reviewed chapter

Clustering Techniques for Land Use Land Cover Classification of Remotely Sensed Images

Written By

Debasish Chakraborty

Submitted: 03 June 2019 Reviewed: 13 August 2019 Published: 29 October 2019

DOI: 10.5772/intechopen.89165

From the Edited Volume

Geographic Information Systems in Geospatial Intelligence

Edited by Rustam B. Rustamov

Chapter metrics overview

1,048 Chapter Downloads

View Full Metrics

Abstract

Image processing is growing fast and persistently. The idea of remotely sensed image clustering is to categorize the image into meaningful land use land cover classes with respect to a particular application. Image clustering is a technique to group an image into units or categories that are homogeneous with respect to one or more characteristics. There are many algorithms and techniques that have been developed to solve image clustering problems, though, none of the method is a general solution. This chapter will highlight the various clustering techniques that bring together the current development on clustering and explores the potentiality of those techniques in extracting earth surface features information from high spatial resolution remotely sensed imageries. It also will provide an insight about the existing mathematical methods and its application to image clustering. Special emphasis will be given on Hölder exponent (HE) and Variance (VAR). HE and VAR are well-established techniques for texture analysis. This chapter will highlight about the Hölder exponent and variance-based clustering method for classifying land use/land cover in high spatial resolution remotely sensed images.

Keywords

  • remote sensing image
  • clustering
  • classification
  • land use
  • land cover
  • features
  • extraction

1. Introduction

High spatial resolution remotely sensed imagery helps to obtain quality and detailed information about the earth’s surface features in conjunction with their geographical associations. The internal changeability within the identical land-use land-cover units augments with the rise of resolution. The augmented changeability diminishes the statistical distinguishability of land-use/land-cover classes in the spectral data space. This reduced distinguishability tends to decrease the accuracies of pixel-based clustering algorithms such as Fuzzy C Means [1], minimum distance classifiers [2] and K-Means [3]. These pixel-based clustering techniques assign a pixel to a region according to the similarities of spectral signature. It considers only one pixel at a time [4]. Spectral signatures are the specific combination of emitted, reflected or absorbed electromagnetic (EM) radiation at varying wavelengths which can uniquely identify an object [4].

Compared to IRS-1A/1B sensors, the spectral resolution of high spatial resolution images is normally relatively poor. Spectral resolution describes a sensor’s ability to identify fine intervals of wavelength. The better the spectral resolution, the finer the channel or band width. Therefore, between spatial and spectral resolution, there is a trade-off. It is mainly true for panchromatic (PAN) images of high spatial resolution, namely CARTOSAT-II 1m and IKONOS 1m. There is a need to consider the spatial relationships between pixel values, also known as the ‘texture’ of the scene objects to classify high-resolution (HR) images owing to the wide difference in the spatial structure in these images. Consequently, multiple texture-based clustering technique namely GLCM [5, 6, 7, 8], Markov random field (MRF) model [5], Gray scale rotation invariant [9] were evolved for clustering remote sensing images having high spatial resolution. Nevertheless, above mentioned methods are appropriate in textured area of HR images. A region is called textured; where the intensity dissimilarity within adjacent pixels is substantial. A region is said to be non-textured, where the intensity dissimilarity among adjacent pixels is insignificant [10, 11]. But texture-based classification techniques failed in non-textured region of high spatial resolution image as much variation is not found in the spatial pattern of those regions of the image [12]. Thus, we can infer from earlier studies that classification of high spatial resolution imageries either by pixel or texture-based algorithm may not yield desired results.

Some more techniques namely watershed approach [13, 14], region-growing approach [4, 15], mean shift approach [16, 17], region merging approach [18] etc. are in use for clustering high spatial resolution remote sensing images. Application of these approaches for clustering of images either leads to under-segmentation or over-segmentation [19, 20]. Structural image indexing approach [21], semi-supervised feature learning approach [22] and multi-scale manner using SVM approach [23] are also found fairly suitable in clustering high resolution images. The imagery of higher resolution includes textured and non-textured areas. Hence, pixel or texture-based algorithm for clustering of high-resolution imagery does not produce expected results. This type of high-resolution imagery clustering research is in the trend. Multi-circular local binary pattern and variance-based method [10] were used separately to cluster high resolution image having textured and non-textured regions. The Multi circular local binary pattern operator has been used here for measuring the spatial structure of the image. But, disadvantage in this strategy is that multi-circular local binary pattern operator is susceptible to noise as it exactly sees the value of the moving window’s central pixel as a limit for computing the spatial structure around the central pixel.

In last one decade the Hölder exponent (HE) has been used for calculating spatial structure of the images [24, 25, 26]. It is also being used for clustering high-resolution images [12]. HE gives an evidence of the spatial structure of the image and is not much influenced by the noise. In addition, spatial structure, contrast of the local image holds considerable property for calculating the texture around the pixel. In this research, high-resolution picture textured and non-textured region is originally segmented using HE and VAR-based method and subsequently separately clustered and non-textured areas. VAR is used to calculate the contrast around the pixel. The suggested method is applied with a 1 m spatial resolution on high resolution IKONOS PAN images.

Advertisement

2. Methods

The suggested high-resolution image ‘P’ clustering technique has three main steps: (i) image transformation, (ii) segmentation and extraction, and (iii) clustering. Initially, every pixel of the image is converted into a degree of texture or non-texture around the pixel. In the second step, using segmented image mask, the transformed image is segmented and non-textured and textured regions are extracted from the initial image. Finally, the two areas obtained are separately clustered.

2.1 Transformation of image

The Hölder Exponent (HE) and VAR are jointly used to convert the image for computing the texture. The HE calculates each pixel of P’s spatial structure. Besides spatial structure, local image contrast also grasps important property for computing the texture around the pixel. In this research, therefore, VAR is used to calculate the contrast around the pixel.

2.1.1 Hölder exponent

Hölder exponent has been used for investigating the texture in high-resolution images [12]. It measures the irregularity in the vicinity. Supremacy of applying Hölder Exponent in HR images are that (i) it can be used as an instrument to calculate each pixel of the image’s spatial structure, (ii) no previous data on the pixel intensity is required and (iii) is not very sensitive to noise [12].

Definition of HE [27]: Let μ be a measure on a set Ω as well as for all x Є Ω, э α(x), such that μ (Br(x)) ∼ rα, for small r. Here Br(x) is circle (2D) of radius i centered on x. Then α (x) is called the HE on x.

A sequence of 15 values of radius r (i.e. 1, √2, √5, 3, √13, 3√2, 5, √29, 2√10, 3√5, 7, √61, 6√2, √85, 7√2) centered on x are used as a scale parameter for calculating HE value around each pixel x in the image [12] and the total number (N) of intersected pixels by the perimeter of series of circles of radius r is considered as a scale parameter for computing VAR value around x [12]. N is computed using Eq. (1).

N=r=1tmrE1

where t is the total number of identified circles, mr is the number of intersected pixels on the perimeter of the radius r circle.

2.1.2 VAR (σ2) for contrast measurement around each pixel of the image

To get the contrast value of (x, y), the neighbor’s σ2 of each pixel (x, y) is calculated over the entire image. Using Eq. (2), the σ2 (x, y) is realized

σ2xy=r=1tj=1mrarjμ2NE2

where arj is the intensity value of pixel (r,j),

μ=r=1tj=1mrarjN

Thus obtained α(x,y) and σ2 (x,y) for each P(x,y). Afterward, these values are used in Eq. (3) to obtain the corresponding pixel value (x,y) in the transformed image T. Each pixel (x,y) of T signifies the degree of texture around that pixel.

Txy=αxy+σ2xy2E3

2.2 Image segmentation and extraction

The image ‘T’ is segmented into textured and non-textured regions based on a threshold value ‘δ’. The pixel value in T below the ‘δ’ is considered to be a non-textured region, whereas greater than or equal to ‘δ’ is considered to be the textured region in the segmented image. Pixels are labeled as zero in non-textured areas, whereas pixels are marked as one in textured areas in the segmented image mask and depicted as follows:

Γxy=1,Txyδ0,Txy<δE4

where T(x,y) and Γ(x,y) represents the pixel value in (x,y) position of the two dimensional transformed image and segmented image respectively and δ represents the threshold value. The δ is calculated by using Eq. (5).

δ=Tmin+TmaxTminKE5

where Tmin and Tmax represents minimum and maximum pixel gray value in T respectively and K is user defined value.

IKONOS PAN sensor image of size 256 × 256 pixels (shown in Figure 1a) is used to achieve the optimum K. The suggested clustering method is also implemented for distinct K values on this image.

Figure 1.

(a) IKONOS image showing vegetation, built-up area, fallow and water body categories, (b) classified image obtained by applying “HE-VAR and PAN” based method on Figure 1a, (c) classified image obtained by applying “MCLBP and VAR” based method on Figure 1a, (d) classified image obtained by applying “proposed classification method” on Figure 1a, (e) IKONOS image showing fallow, water bodies, vegetation and built-up area categories, (f) classified image obtained by applying “HE-VAR and PAN” based method on Figure 1e, (g) classified images obtained by applying “MCLBP and VAR” based method on Figure 1e, (h) Classified images obtained by applying “proposed classification method” on Figure 1e.

The segmented image is subsequently used to obtain the textured and non-textured region from the initial image P. This process’s mathematical representation is shown as follows:

R1xy=Pxy,Γxy=00,Γxy0E6
R2xy=Pxy,Γxy=10,Γxy1E7

where P, Γ, R1 and R2 indicates original image, segmented image, extracted non-textured region from original image P and extracted textured region from original image P respectively.

2.3 Clustering

Initially, a threshold is used to segment the transformed image into textured and non-textured region. Afterward, the original image is extracted into textured and non-textured regions using the segmented image mask and clustered independently. The extracted textured region (R2) is clustered by means of ISODATA clustering algorithm [28] considering HE, VAR and intensity values of individual pixel of textured area. The clustering algorithm of ISODATA is less computational, easy and non-supervisory. Whereas the non-textured area (R1) of the image is categorized using the clustering algorithm of ISODATA. In the event of non-textured region, the individual pixel HE and VAR value is not regarded for classification as there is no important variation in texture between classes. The classified outputs of the non-textured and textured region are subsequently produced separately and mixed together to obtain the final classified image.

This research uses “HE-VAR and PAN” and “MCLBP and VAR” based clustering technique to show the power of the suggested clustering technique. The technique based on “HE-VAR and PAN” clusters the entire image using the HE, VAR and intensity of each pixel of the IKONOS PAN image. The suggested technique of clustering is then contrasted with the outcomes of the clustering method based on “HE-VAR and PAN” and “MCLBP and VAR” to demonstrate the strength of the suggested technique of clustering.

Advertisement

3. Results and discussion

The projected clustering method imagines threshold δ to get the segmented image mask from the transformed image. The threshold is computed using a constant ‘K’. In this study, proposed clustering procedure is implemented on IKONOS PAN image with spatial resolution 1 m for ‘K’ values between 3 and 7 and subsequently, classification rate is measured for these ‘K’ values using the ground truth data. The classification accuracy with different ‘K’ is shown in Figure 2. The ‘K’ affects the accuracy in classifying High spatial resolution images considerably as shown in Figure 2. For computing texture, a suitable choice of ‘K’ is important. In this study, superlative performance in high-resolution image classification was accomplished with K = 5. The optimum K is discovered based on Figure 1a and is also implemented in the classification of Figure 1e in addition to other images and found classification accuracy is more than 88%. Thus, from the present study, we can infer that the same K value is suitable for most images.

Figure 2.

Classification accuracy as a function of K.

The Proposed clustering method, “MCLBP and VAR” based method and HE-VAR and PAN” based method were applied on two different 1 m PAN (IKONOS) images (size 256 × 256 pixels) covering (i) vegetation, (ii) built-up area, (iii) water bodies, and (iv) fallow (shown in Figure 1a,e). Texture is observable in in Figure 1a,e. The results of proposed method are then compared with the results obtained from the analysis based on “HE-VAR and PAN” and “MCLBP and VAR” respectively.

Figure 1fh shows the classification outcomes of the methods “HE-VAR and PAN,” “MCLBP and VAR” and “Proposed classification” after proceeding to the second IKONOS image respectively. Figure 1bd shows the classification outcomes of the methods “HE-VAR and PAN,” “MCLBP and VAR” and “Proposed classification” after proceeding to the first IKONOS image respectively. Classified images recognize varied features in Figure 1bd,fh. From the results, it is evident that the method based on “MCLBP and VAR” gives less heterogeneous segments than the method based on “HE-VAR and PAN,” while the method based on “Proposed classification method” provides more homogeneous segments with distinct classes than the method based on “MCLBP and VAR.”

The ground truth data is collected using GPS equipment for the class vegetation, built-up area, fallow and water body of sample size of 656, 519, 577 and 462 square meters respectively. Afterward, ArcGIS software is used to transfer the ground truth data into vector data. Subsequently, by overlaying the ground truth information distinctly on the results acquired from both IKONOS images (Figure 1a,e) adopting methods such as “HE-VAR and PAN,” “MCLBP and VAR” and “Proposed clustering,” the classification accuracies for each strategy are shown by confusion matrix. The confusion matrices (Table 1) calculated for Figure 1bd showed that the precision of classification of vegetation, built-up area, fallow and water bodies is (73, 69, 59 and 87% respectively) based on the ‘HE-VAR and PAN’ technique and (79, 71, 68 and 89% respectively) based on the ‘MCLBP and VAR ‘technique, whereas (91, 86, 85 and 94% respectively) by the “Proposed clustering” method. Table 2 demonstrates the confusion matrices calculated for Figure 1fh showed that the precision of classification of vegetation, built-up area, fallow and water bodies is (73, 74, 66 and 88% respectively) based on the ‘HE-VAR and PAN ‘technique and (78, 76, 68 and 89% respectively) based on the ‘MCLBP and VAR ‘technique whereas (90, 87, 86 and 93% respectively) by the “Proposed clustering” method.

Classes derived from satellites
Classification methodGrand observed classVegetationBuilt-up areaFallowWater bodyRow totalClassification accuracy (%)C (%)O (%)
HE-VAR and PANVegetation2513512730572.7527.2415.65
Built-up area1816314520069.0730.9315.68
Fallow53231441423459.2640.7437.04
Water body23157317728887.1912.8154.67
Column total3452362432031027
MCLBP and VARVegetation2723111433278.8418.2613.33
Built-up area171687320571.2528.8111.41
Fallow43211661522368.3331.6932.51
Water body13165918125889.2810.8343.35
Column total3452362432031027
Proposed methodVegetation313177334090.859.277.82
Built-up area72048222186.2813.567.20
Fallow197206723984.7715.2313.58
Water body682219122794.115.9117.73
Column total3452362432031027

Table 1.

The confusion matrices showing the classification accuracy obtained by applying “HE-VAR and PAN”, “MCLBP and VAR” and “Proposed” methods separately on IKONOS image shown in Figure 1a.

C: Commission error, O: Omission error.

Classes derived from satellites
Classification methodGrand observed classVegetationBuilt-up areaFallowWater bodyRow totalClassification accuracy (%)C (%)O (%)
HE-VAR and PANVegetation22815431630273.326.6823.79
Built-up area3220914526073.726.1418.02
Fallow27272211128666.333.8319.46
Water body24325622733987.612.3543.24
Column total3112833342591187
MCLBP and VARVegetation2411338930177.622.5120.58
Built-up area2921513726475.8324.0217.31
Fallow26262281129168.3631.7418.86
Water body15295523233189.410.4238.22
Column total3112833342591187
Proposed methodVegetation281714630890.49.648.68
Built-up area32449526186.713.786.0
Fallow1515287732485.814.0711.07
Water body12172424129493.26.9420.46
Column total3112833342591187

Table 2.

The confusion matrices showing the classification accuracy obtained by applying “HE-VAR and PAN”, “MCLBP and VAR” and “Proposed” methods separately on IKONOS image shown in Figure 1e.

C: Commission error, O: Omission error.

The categorized result for Figure 1a,e shows that the “HE-VAR and PAN” method under segment as a result (i) fallow assorted with water bodies shown in Figure 1b,f, (ii) built-up region assorted with fallow and vegetation shown in Figure 1f, (iii) vegetation assorted with water bodies shown in Figure 1b,f, (iv) fallow assorted with built-up region shown in Figure 1b. This incoherence decreases vegetation, fallow, water bodies and built-up area classification precision as shown in Tables 1 and 2. The technique based on “MCLBP and VAR” somehow overcomes these inconsistencies. It is discovered that, as shown in Figure 1c,g, the superposition of fallow, water body, vegetation region becomes less. In addition, decreased inconsistencies improve the accuracy of the classification of fallow, water body and vegetation regions (see Tables 1 and 2).

“HE-VAR and PAN” based method classifies water bodies and fallow areas as a single class (Figure 1b,f) since the texture patterns of these two areas does not show much difference in high resolution imageries as shown in Figure 1a,e. “MCLBP and VAR” based technique demonstrates improvement in classifying the fallow areas and water bodies which is observable in Figure 1g. But this method could not extract non-textured region appropriately form Figure 1a since MCLBP is sensitive to noise. Therefore “MCLBP and VAR” based method could not discriminate appropriately fallow areas and water bodies in Figure 1a as visible in Figure 1c. HE is not as much of sensitive to noise therefore the proposed technique partitions the image into textured and non-textured regions noticeably which in turn helps in classifying the fallow and water bodies as shown in Figure 1d.

The proposed clustering method is applied further on a 1 m PAN (IKONOS) image (Figure 3a) of (i) urban woodland, (ii) building, (iii) water bodies, and (iv) fallow to show the robustness and validity of the method in classifying land use area. The method satisfactorily discriminate urban woodland, building, fallow and water bodies as shown in Figure 3b. The algorithm also implemented on two extra 1 m PAN (IKONOS) images: (i) Figure 4a of fallow, vegetation, built-up area and bare land and (ii) Figure 4c of water, vegetation, fallow and built-up area. The findings (Figure 4b,d) show that vegetation, fallow, built-up region, bare soil and water bodies are satisfactorily discriminated against by the algorithm.

Figure 3.

(a) IKONOS image showing urban woodland, building, water body and fallow categories, (b) classified image obtained by applying “proposed classification method” on Figure 3a.

Figure 4.

(a) IKONOS image showing fallow, built-up area, vegetation and bare soil categories, (b) classified image obtained by applying “proposed classification method” on Figure 4a, (c) IKONOS image showing vegetation, fallow, built-up area and water bodies categories, (d) classified image obtained by applying “proposed classification method” on Figure 4c.

Advertisement

4. Conclusion

In the present study, the spatial structure of local image texture is computed using HE. The contrast around the pixel is measured using VAR. Afterward, the image is transformed using HE and VAR together for measuring the texture. A threshold δ is used to extract textured and non-textured region from the image. The classification algorithm ISODATA is used to classify the textured region taking into account HE, VAR and intensity values of the textured area’s individual pixels. Whereas ISODATA clustering algorithm classifies the extracted non-textured region of the image. The HE and VAR value of individual pixels is not regarded for classification in the event of non-textured region. From the research outcomes, it is discovered that the suggested technique is helpful to extract earth surface characteristics from complicated remote sensing images that contain both textured and non-textured areas. Moreover, it can be considered as an intuitively appealing and unsupervised clustering algorithm for extracting features from remotely sensed images. As a result, the method is potentially useful to extract earth surface features by clustering high spatial resolution panchromatic images more efficiently.

Advertisement

Acknowledgments

The author sincerely thanks the Director, NRSC, Hyderabad, India and CGM, RCs, NRSC, Hyderabad, India for their support. The author is also grateful to the former GM, RRSC-East for support.

References

  1. 1. Bezdek JC, Ehrlich R, Full W. FCM: The Fuzzy C-Means clustering algorithm. Computers and Geosciences. 1984;10:191-203
  2. 2. Richards JA. Remote Sensing Digital Image Analysis: An Introduction. Berlin, Heidelberg: Springer-Verlag; 1995. pp. 265-290
  3. 3. Hartigan JA, Wong MA. A K-means clustering algorithm. Applied Statistics. 1979;28(1):100-108
  4. 4. Chakraborty D, Sen GK, Hazra S. Image Segmentation Techniques. Germany: LAP LAMBERT Academic Publishing; 2012. pp. 1-128
  5. 5. Clausi DA, Yue B. Texture segmentation comparison using grey level co-occurrence probabilities and Markov random fields. In: Proceedings of the 17th International Conference on Pattern Recognition (ICPR’04); 2004. Available from: http://ieeexplore.ieee.org/stamp/
  6. 6. Haralick RM, Shanmugan K, Dinstein I. Textural features for image classification. IEEE Transactions on Systems, Man, and Cybernetics. 1973;6:610-621
  7. 7. Tsai F, Chou MJ. Texture augmented analysis of high resolution satellite imagery in detecting invasive plant species. Journal of the Chinese Institute of Engineers. 2006;29(4):581-592
  8. 8. Tsai F, Chou MJ, Wang HH. Texture analysis of high resolution satellite imagery for mapping invasive plants. In: International Geoscience and Remote Sensing Symposium; 2005. pp. 43024-43027
  9. 9. Klemas VV. Remote sensing of coastal resources and environment. Environmental Research, Engineering and Management. 2009;2(48):11-18
  10. 10. Chakraborty D, Chowdhary VM, Dutta D, Sharma JR. Classification of high spatial resolution image using multi circular local binary pattern and variance. International Journal of Electronics and Computer Engineering. 2013;4(6):1648-1653
  11. 11. Weng Q. Remote sensing of impervious surfaces in the urban areas: Requirements, methods, and trends. Remote Sensing of Environment. 2012;117:34-49
  12. 12. Chakraborty D, Sen GK, Hazra S. High-resolution satellite image segmentation using H¨older exponents. Journal of Earth System Science. 2009;118(5):609-617
  13. 13. Mathivanan B, Selvarajan S. High spatial resolution remote sensing image segmentation using marker based watershed algorithm. Journal of Academia and Industrial Research. 2012;1(5):257-260
  14. 14. Wang Z, Zhao S, Chen X. Watershed segmentation of high-resolution remotely sensed data. In: Proceedings of International Symposium on Remote Sensing & 20th Anniversary of the Korean Society of Remote Sensing; 2004. pp. 107-109
  15. 15. Carleer AP, Debeir O, Wolff E. Assessment of very high spatial resolution satellite image segmentations. Photogrammetric Engineering and Remote Sensing. 2005;71(11):1285-1294
  16. 16. Chakraborty D, Sen GK, Hazra S, Jeyaram A. Clustering for high resolution monochrome satellite image segmentation. International Journal of Geoinformatics. 2008;4(1):1-9
  17. 17. Su T, Li H, Zhang S, Li Y. Image segmentation using mean shift for extracting croplands from high-resolution remote sensing imagery. Remote Sensing Letters. 2015;6(12):952-961
  18. 18. Zhang X, Xiao P, Feng X, Wang J, Wang Z. Hybrid region merging method for segmentation of high-resolution remote sensing images. ISPRS Journal of Photogrammetry and Remote Sensing. 2014;98:19-28
  19. 19. Chen B, Qiu F, Wu B, Du H. Image segmentation based on constrained spectral variance difference and edge penalty. Remote Sensing. 2015;7:5980-6004
  20. 20. Wang C, Shi AY, Wang X, Wu FM, Huang FC, Xu LZ. A novel multi-scale segmentation algorithm for high resolution remote sensing images based on wavelet transform and improved JSEG algorithm. Optik - International Journal for Light and Electron Optics. 2014;125(19):5588-5595
  21. 21. Xia GS, Yang W, Delon J, Gousseau Y, Sun H, Maître H. Structural high-resolution satellite image indexing. In: ISPRS TC VII Symposium-100 Years ISPRS, Vol. 38; 2010. pp. 298-303
  22. 22. Yang W, Yin X, Xia GS. Learning high-level features for satellite image classification with limited labeled samples. IEEE Transactions on Geoscience and Remote Sensing. 2015;53(8):4472-4482
  23. 23. Huang X, Zhang L. An SVM ensemble approach combining spectral, structural, and semantic features for the classification of high-resolution remotely sensed imagery. IEEE Transactions on Geoscience and Remote Sensing. 2013;51(1):257-272
  24. 24. Lucieer A, Stein A, Fisher P. Multivariate texture-based segmentation of remotely sensed imagery for extraction of objects and their uncertainty. International Journal of Remote Sensing. 2005;26(14):2917-2936
  25. 25. Malladi RK, Kasilingam D, Costa AH. Speckle filtering of SAR images using Hölder regularity analysis of the sparse code. In: IEEE International Geoscience and Remote Sensing Symposium; 2003. pp. 63998-64000
  26. 26. Tahiri AM, Farssi SM, Touzani A. Textures in Images Classification Using a Multifractal Approach. IEEE SITIS. 2005. Available from: http://www.u-bourgogne.fr/SITIS/05/download/Proceedings/Files/f138.pdf
  27. 27. Bourissou A, Pham K, Levy-Vehel J. A multifractal approach for terrain characterization and classification on SAR images. IGARSS. 1994;3:1609-1611
  28. 28. Jain AK, Murty MN, Flynn PJ. Data clustering: A review. ACM Computing Surveys. 1999;31(3):264-323

Written By

Debasish Chakraborty

Submitted: 03 June 2019 Reviewed: 13 August 2019 Published: 29 October 2019