Spectral Clustering with Eigenvalue Similarity Metric Method for POL-SAR Image Segmentation of Land Cover ()

Shuiping Gou^{}, Debo Li^{}, Dong Hai^{}, Wenshuai Chen^{}, Fangfang Du^{}, Licheng Jiao^{}

Xidian University, Xi’an, China.

**DOI: **10.4236/jgis.2018.101007
PDF HTML XML
949
Downloads
2,204
Views
Citations

Xidian University, Xi’an, China.

A simple and fast approach based on eigenvalue similarity metric for Polarimetric SAR image segmentation of Land Cover is proposed in this paper. The approach uses eigenvalues of the coherency matrix as to construct similarity metric of clustering algorithm to segment SAR image. The Mahalanobis distance is used to metric pairwise similarity between pixels to avoid the manual scale parameter tuning in previous spectral clustering method. Furthermore, the spatial coherence constraints and spectral clustering ensemble are employed to stabilize and improve the segmentation performance. All experiments are carried out on three sets of Polarimetric SAR data. The experimental results show that the proposed method is superior to other comparison methods.

Keywords

Polarimetric Synthetic Aperture Radar, Eigenvalue, Mahalanobis Distance, Spectral Clustering, Image Segmentation

Share and Cite:

Gou, S. , Li, D. , Hai, D. , Chen, W. , Du, F. and Jiao, L. (2018) Spectral Clustering with Eigenvalue Similarity Metric Method for POL-SAR Image Segmentation of Land Cover. *Journal of Geographic Information System*, **10**, 150-164. doi: 10.4236/jgis.2018.101007.

1. Introduction

The fully-polarimetric synthetic aperture radar (SAR) [1] [2] [3] has the ability to provide information in four channels HH, HV, VH and VV, and contains complete polarization information of electromagnetic waves’ effect on surface. We can obtain the information of land cover by analysis and interpretation of the polarization information. Segmentation for land cover is one of basic issues and also important applications.

The segmentation process is based on the choice of features and classifier. Selecting good features can obtain better segmentation result than only improving classifier. In the existing segmentation of polarimetric SAR, what is generally used as features is polarimetric information, and the texture or gray information of image [4] [5] . The prior work has shown that the eigenvalues of coherency matrix include rich polarimetric information with good divisibility [6] . At the same time, the eigenvalue, which is obeying Gaussian distribution approximately, is easy to be metric.

The existing classifier can be divided into the supervised and the unsupervised [7] [8] . Compared with the unsupervised clustering, the supervised needs partially landcover labels, thus can often get better segmentation result [9] [10] . However, due to the landcover label of polarimetric SAR image is difficult to obtain, more researchers focus on the unsupervised cluster methods [11] - [19] . Cloude et al. have previously used the threshold of scattering entropy, scattering angle and inverse entropy to classify [11] [14] [15] . Freeman et al. have extracted three scattering powers and classified according to the proportion [17] . Lee et al. have proposed the maximum likelihood classifier based on complex Wishart distribution [18] . The above unsupervised clustering methods are faced up the choice of threshold that needs much artificial experience which is cost. So we choose the real unsupervised clustering method as classifier.

Clustering usually means to group in accordance with the similarity between objects. The distance is the most common similarity metric, which reflects the similarity between objects by measuring the difference of the objects. In practical applications, the choice of distance depends on the characteristics of the object, and it is generally applied to cluster as similarity metric. In the used POL-SAR segmentation, according to the coherency matrix which is obeying the Wishart distribution, Anfinsen [20] and Ersahin [21] have chosen the Wishart distance with Gaussian kernel as the similarity metric, and applied it to spectral clustering. This kind of special distribution of the T-matrix limits the construction of similarity metric. And there are two problems calculating the similarity with Gaussian kernel between pixels: 1) the scale parameter of the Gaussian kernel σ needs to be manually set precisely in accordance with the experience, and the single scale parameter σ cannot capture category distribution information of multiple scales data well; 2) huge exponentiation consumption.

In our method, firstly, we choose the eigenvalues of coherency matrix as the input features, which include the essential information of coherency matrix and represent the intensity of scattering. Secondly, we apply Mahalanobis distance as the similarity metric by studying the statistical properties of eigenvalues. And taking into account the neighborhood information of image, consistency constraints will be applied to the similarity metric. Thirdly, the above similarity metric is applied to spectral clustering algorithm to complete the segmentation. At last, in order to improve and stabilize the segmentation results, the strategy of cluster ensemble is used.

2. Feature Extraction and Its Similarity Metric

2.1. Polarmetric Features Analysis

The full-polarimetric SAR data can be expressed by complex scattering matrix $S$ :

$S=\left[\begin{array}{cc}{S}_{hh}& {S}_{hv}\\ {S}_{vh}& {S}_{vv}\end{array}\right]$ (1)

where, $h$ and $v$ represent horizontal and vertical polarization modes, respectively. It is commonly assumed that natural targets exhibit reciprocity, ${S}_{hv}={S}_{vh}$ . The above scattering matrix also can then be expressed as the scattering vector $k$ .

$k=\frac{1}{\sqrt{2}}\left[{S}_{hh}+{S}_{vv},{S}_{hh}-{S}_{vv},2{S}_{hv}\right]$ (2)

In order to better explain the physical meaning of the scattering process, the coherency matrix T is used:

$T=k\cdot {k}^{*}$ (3)

The coherency matrix is a Hermit matrix, $T={T}^{\ast}$ , whose size is 3 × 3.

In order to make better use of the polarization scattering matrix to reveal the physical mechanism, polarization data usually is broken down into different components by decomposition [14] . In [11] , Cloude et al. have put forward Cloude decomposition which has very important significance in POL SAR data processing. The two parameters, scattering entropy and the scattering angle, obtained from decomposition are widely used in image segmentation [18] .

Cloude decomposition:

$\left[T\right]=\left[{U}_{3}\right]\left[\begin{array}{ccc}{\lambda}_{1}& 0& 0\\ 0& {\lambda}_{2}& 0\\ 0& 0& {\lambda}_{3}\end{array}\right]{\left[{U}_{3}\right]}^{*\text{T}}$ (4)

where ${\lambda}_{i}$ is eigenvalue, $\left[{U}_{i}\right]$ is eigenvector corresponding to eigenvalue ${\lambda}_{i}$ , $i=1,2,3$ . Each eigenvector represents a scattering mechanism, and the corresponding eigenvalue represents the intensity of the scattering mechanisms.

The eigenvalues and eigenvectors resulting from Cloude decomposition have been concerned and studied [22] [23] [24] [25] . Carlos Lopez-Martinez has made an in-depth analysis of the probability density function of the sample eigenvalues of the covariance or coherency matrix and proposed that Gaussian scattering assumption is valid to the sample eigenvalues distribution of homogeneous distributed scatters [22] [24] . In [22] , the Gaussian Mixture Modes have been used in inhomogeneous areas. Based on different areas obeying Gaussian distribution with different parameter, expectation maximization algorithm has been applied to the segmentation of POL-SAR data.

So we conclude that the eigenvalues of coherency matrix include rich polarization information. And Gaussian distribution of eigenvalues makes it more convenient to measure than Wishart distribution of coherency matrix.

2.2. Similarity Metric Based on Eigenvalue Analysis

As described in the reference [22] , we analyze the distribution characteristics of eigenvalue. Figure 1(a) marked seven areas A-F, comes from a part of the POL-SAR data set Flevoland. 50 sample points are randomly selected from every area, respectively. Then, we make three eigenvalues of every point as coordinate and show them as is seen from Figure 1(b). We can see that the eigenvalues of different class have divisibility substantially. Then, we apply Gaussian Mixture Modes to simulate eigenvalues of every area, as is seen from Figure 1(c). We can conclude that eigenvalues are approximate Gaussian distribution with different parameter.

Euclidean distance is the most widely used similarity metric, whose characteristics are as follows: 1) the ranges of different features (Table 1) are ignored. As

Figure 1. Distribution of different areas eigenvalues.

Table 1. Comparison of range of three eigenvalues.

result, the Euclidean distance between two points depend on the feature ${\lambda}_{3}$ largely. 2) Without considering the correlation between features, Euclidean distance treats features equally and only integrates the difference of each feature between two points. 3) Euclidean distance is applicable to the data obeying strictly Gaussian distribution. However, our data is not strictly Gaussian distribution, just like the area A shown in Figure 1(c).

In order to solve the above problems of Euclidean distance, Indian statistician P. C. Mahalanobis has proposed Mahalanobis distance based Multivariate Statistics. It is an effective method of similarity metric between unknown sample sets, and is also called covariance distance. With respect to the Euclidean distance, Mahalanobis distance has the following advantages: 1) Mahalanobis distance is normalized distance of non-uniform distribution in the Euclidean space, balancing the ranges of different features. 2) Mahalanobis distance is based on the distribution of features in the entire space, therefore, to better describe the similarity between two points. 3) Mahalanobis distance is applicable to the data obeying Gaussian distribution approximately [26] .

${S}_{\text{Euclidean}}=\sqrt{{\left(x-y\right)}^{\text{T}}\cdot \left(x-y\right)}$ , (5)

${S}_{\text{Ma}}=\sqrt{{\left(x-y\right)}^{\text{T}}\cdot {C}^{-1}\cdot \left(x-y\right)}$ , (6)

where, $x,y$ are features of two points. $C$ is covariance matrix, varying with input data. Thus, the similarity metric with Mahalanobis distance is adaptive.

2.3. Similarity Metric with Spatial Consistency Constraint

In the process of segmentation and classification, the probability of pixels in the image and its neighborhood having the same class attributes is large, called spatial coherence constraints. So, we choose the similarity metric with spatial coherence constraints [27] .

In summary, the final used affinity matrix is:

$S+\alpha \stackrel{\xaf}{S}$ (7)

where,

${S}_{ij}=\sqrt{{\left({x}_{i}-{x}_{j}\right)}^{\text{T}}\cdot {C}^{-1}\cdot \left({x}_{i}-{x}_{j}\right)}$ , (8)

${\stackrel{\xaf}{S}}_{ij}=\frac{1}{{N}_{R}}{\displaystyle \underset{{x}_{r}\in {N}_{k}}{\sum}\sqrt{{\left({x}_{i}-{x}_{r}\right)}^{\text{T}}\cdot {C}^{-1}\cdot \left({x}_{i}-{x}_{r}\right)}}$ (9)

where, $C$ is covariance matrix, ${x}_{i},{x}_{j}$ are the features of the ith, jth image pixels. ${N}_{k}$ are the pixels whose center is ${x}_{j}$ and neighborhood window is $k\times k$ . ${N}_{R}$ is the number of pixels included in window.

As described in [28] , this method of parameter $\alpha $ is not sensitive. When $\alpha \ge 3$ , the result of algorithm is remained. And, when the size of the neighborhood window $k\le 15$ , the result is improved gradually. When $k>15$ , the result is bad gradually. Specific to each image, the k is decided by the texture.

With above similarity metric, grouping data is to complete clustering. However, some of the commonly used statistical-based clustering algorithm such as EM, demands obedience distribution, and is sensitive to the initialization. At the same time, when Gaussian fitting, the statistical properties of mixed-pixel is unstable. So EM algorithm is not suitable for the eigenvalues, just like the results displayed in Figure 2. As result, we choose the spectral clustering whose distribution of data is not a requirement.

2.4. Spectral Clustering Spatial Consistency

Spectral clustering is a typical clustering algorithm based similarity metric. Spectral clustering algorithm is no longer required a convex structure of the data to ensure a good result, and which is also a discriminant method. Instead of making assumptions of the global structure of data, spectral clustering algorithm firstly collects local information to indicate the possibilities of whether two points belonging to the same class, then makes the global decision based on a clustering criterion to divide all the data points into irrelevant sets.

Spectral clustering has good clustering results, but for larger POLSAR data, the application of classic spectral clustering algorithm has been limited [29] . Many fast spectral clustering algorithms are proposed [30] [31] [32] [33] [34] . In [30] , Fowlkes et al. have put forward Nyström algorithm which is simple, effective, and greatly reduce the computational complexity. So we choose Nyström to cluster. Nyström is a digital approximation technique of solving the problem of

Figure 2. The flow chart of our method.

the integral characteristic function. The method first randomly select a small part of sample from all of the samples as representative points to solve the characteristic problem, and then extend eigenvector to the similarity matrix for the entire sample set.

The main steps of Nyström algorithm are as follows:

Step 1. Randomly selected m sample points as sample subset;

Step 2. Form the affinity matrix of the subset $W\in {R}^{n\times n}$ , where ${W}_{ij}=\mathrm{exp}\left(-{\Vert {x}_{i}-{x}_{j}\Vert}^{2}/2{\sigma}^{2}\right),\text{\hspace{0.17em}}i\ne j,\text{\hspace{0.17em}}{W}_{ii}=0$ ;

Step 3. Eigendecompose W, obtain the eigenvalues and the corresponding eigenvectors of W, then extrapolate eigenvectors of the entire similarity matrix;

Step 4. Clustering first n-dimensional eigenvectors into n clustering via k-means, as the final segmentation results.

Where, $W$ means similarity matrix between sample points to be cluster, and it contains all the information required to cluster. In our method, $W$ is conducted as $S+\alpha \stackrel{\xaf}{S}$ .

2.5. Cluster Ensemble

The clustering ensemble is a final division of multiple clustering results of given task, and the division has better robustness, novelty and stability. The key issue is how to obtain better clustering results based on combinations of different cluster results membership, also means the construction and choice of consensus function.

Consensus function gives multiple clustering results a final division. In [29] , the researcher has introduced three consensus functions: cluster-based similarity partitioning algorithm (CSPA), hyperGraph partitioning algorithm (HGPA), meta-Clustering algorithm (MCLA). All of them approach the problem by first transforming the set of clustering into a hypergraph representation, and obtaining hypergraph minimum cut. Among of them, the complexity of MCLA is $O\left(n{k}^{2}{r}^{2}\right)$ , varying linearly with the number of samples. MCLA has superiority in complexity and the integrated quality, so we choose MCLA as consensus function.

Nyström algorithm can effectively reduce the computational complexity. However, clustering results are instable as a result of the random sampling. So we make use of cluster ensembles to keep stable segmentation results. Apply Nyström algorithm for k times, and every time random sampling the same amount of sample to obtain clustering labels $\left\{labe{l}_{1},labe{l}_{2},\cdots ,labe{l}_{k}\right\}$ . And map the labels into the final result with MCLA.

3. Segmentation Algorithm and Experiment Results

3.1. Algorithm Steps of Eigenvalue Similarity Metric Based Spectral Clustering Method

Our algorithm process is divided into three steps: pretreatment, similarity metric, spectral clustering ensemble. Pretreatment: refined Lee filter with window 7 × 7, and Cloude decompose to obtain eigenvalues as input features. Similarity metric: construct similarity matrix with Mahalanobis distance. Spectral clustering ensemble: spectral cluster for several times and ensemble with MCLA. The flow chart of eigenvalue similarity metric based spectral clustering is as shown in Figure 2.

3.2. Experiment Data

1) Flevoland Data Set：It is NASA/JPL AIRSAR L-Band POLSAR dataset of Flevoland, the Netherlands, which has the size of 1024 × 750 pixels. The pixel size is 6.6 m in the slant range direction and 12.10 m in the azimuth direction. In Figure 3(a), the image is shown with color composed by Pauli matrix representation: red for $\left|HH-VV\right|$ , green for $\left|HV\right|+\left|VH\right|$ and blue for $\left|HH+VV\right|$ . The map shows an agricultural area, covered with different crops and water.

2) San Francisco Data Set：It is the fully polarimetric L-band airborne SAR data acquired with the AIRSAR sensor of the NASA/JPL at the test site of San Francisco bay, which has a mixed scene of urban, vegetation and ocean. The original data has the size of 1024 × 900 pixels, and experimental data is the size of 800 × 500. In Figure 3(b), the image is shown with color composed by Pauli matrix representation: red for $\left|HH-VV\right|$ , green for $\left|HV\right|+\left|VH\right|$ and blue for $\left|HH+VV\right|$ .

3.3. Feature Analysis

For clustering, the divisibility between two categories depends on the distance between them. So we make a comparison of Euclidean distance and Mahalanobis distance of each of the two categories in Figure 4(a) and Figure 4(b). We can see from ordinate that Mahalanobis distance increase the distance between the two similar classes and improve the divisibility. Simultaneously, the problem that the distances of each of the two categories are small, is to improve (In Figure 4(a), each color point is crossover.).

And we also can see the results of Figure 4(a) and Figure 4(b), for the easy mixed categories (A blue, G brown), although the Mahalanobis distance can extend

Figure 3. Experiment Data. (a) Flevoland data (b) San Francisco data.

Figure 4. A-G class were randomly taken 50 points, respectively calculate the distance between each class and all classes (including itself). Each class has its own color, which is consistent with the color of GroundTruth. (a) Euclidean distance. (b) Mahalanobis distance. (c) Pauli RGB composite and the areas marking. (d) Groundtruth. (e) The result of spectral clustering with Euclidean distance. (f) The result of spectral clustering with Mahalanobis distance.

the distance between two categories, but still not enough to separate every point.

3.4. Segmentation Result

In order to better demonstrate the effectiveness of our method, we choose the contrast algorithms: 1) H/a/A_Wishart [19] ; 2) spectral clustering_Wishart proposed by Anfinsen et al. in [20] ; 3) our method with Euclidean distance.

a) OVER ALL Flevoland Data Set

For image Flevoland, in our method,the number of random sample is 70, the size of neighborhood window $k=3$ because this image doesn’t has fine texture but has small blocks, $\alpha =3$ , the number of classes is 7. The number of classes of the algorithm (spectral clustering_Wishart) is also 7, and the algorithm (H/a/A_Wishart) is fixed 16 then mergered to be 7 manually.

The ground truth does not provide a label for each pixel of the entire image so the accuracy calculation is limited to only those pixels where the ground-truth provides a label. Partial ground-truth map is shown in Figure 5(a). The overall segmentation accuracy of four methods is shown in Table 2, and segmentation maps are shown in Figures 5(b)-(e). Overall segmentation accuracy $\stackrel{\xaf}{P}$ is defined

$\stackrel{\xaf}{P}=\frac{1}{N}{\displaystyle \underset{i=1}{\overset{K}{\sum}}{\displaystyle \underset{j=1,i\ne j}{\overset{K}{\sum}}Correct\left(i,j\right)}}$ (10)

where Correct is the number of pixels emergimg both in the ground truth and classifying result for each category terrain object, N is the total number of pixels, and K is the category number of terrain, i, j is the pixel from a kind of category.

From Table 2 and Figure 5 it is seen that the performance of our method with Mahalanobis distance is much better than Euclidean distance. For the edge of region, our method is worse than others as a result of application of spatial information, and other methods make use of Wishart iteration or segmentation to keep clear edge. For the region consistence, our method is better as a result of

Figure 5. Comparing experimental results of Flevoland data. (a) Partial ground-truth map. (b) H/a/A_Wishart. (c) Spectral clustering_Wishart. (d) Our method with Euclidean distance. (e) Our method with Mahalanobis distance.

Table 2. The overall segmentation accuracy comparison of Flevoland area for method.

application of spatial information and Cluster ensembles. In addition, H/a/ A_Wishart and spectral clustering_Wishart have the similar results, but the latter can converge faster.

b) PARTIAL Flevoland Data Set

In order to improve the credibility of the contrast algorithm, and contrast with reference [21] , we choose the same image (Figure 6(a)) which is retrieved from Figure 3(a). And GroundTruth also come from reference [20] . As described in [20] , the H/a/A_Wishart (0.68) and spectral clustering_Wishart (0.67) have the similar results, and its algorithm (0.75) outperform the H/a/A_Wishart by 7.1%. Though our method (0.745) outperform the H/a/A_Wishart by 6.5%, our method has less complex process and fewer parameter.

c) San Francisco Data Set

In order to demonstrate the robustness of our method, the San Francisco bay data is tested by four algorithms. The number of random sample is 70, $k=3$ , $\alpha =1$ , the number of classes is 3. The contrast algorithms results are mergered to be 3 manually. As can be seen from the two marked details, our method is superior to other methods in the shape. At the same time, ocean, city and forest are classified clearly, and the forest at the upper left corner has good regional consistency. The overall segmentation accuracy of four methods is shown in Table 3, and segmentation maps are shown in Figures 7(a)-(d).

Figure 6. Comparing experimental results of partial Flevoland data. (a) Pauli RGB composite. (b) Partial ground-truth map. (c) H/a/A_Wishart from the reference [19] . (d) Spectral clustering_Wishart from the reference [20] . (e) The algorithm from the reference [21] . (f) Our method with Mahalanobis distance.

Table 3. The segmentation accuracy comparison of marked area of San Francisco for methods.

Figure 7. Segmentation results of San Francisco data. (a) H/a/A_Wishart. (b) Spectral clustering_Wishart. (c) Our method with Euclidean distance. (d) Our method with Mahalanobis distance.

In sum up, our method have good segmentation performance. The main advantages of this method are simple, fast and effective. In term of running time, for the image of Flevoland, the Nyström algorithm needs 40 s when we choose the sample number as 70. We choose the ensemble time N as 3, so the running time of the entire program is about 2 minutes. The affect on the images of San Francisco bay and Xi’an city is small by random sampling, so the process of cluster ensemble can be bypassed. And the contrast algorithms need the process of Wishart iteration, which consumes time. In the term of validity, our method can guarantee the accuracy of the overall segmentation, at the same time keep the details on good results.

At the same time, we can see the method has good robustness from the above three images. This is because the eigenvalue expresses the main information of T matrix.

4. Conclusion

This paper introduces an approach for segmentation of the POLSAR data based on eigenvalue similarity metric. From the scientific and application point of view, it is a new approach of data processing. With analysis of the characteristic eigenvalue, we propose a new construction method of similarity metric. As a result, our method reduces the complexity of the spectral clustering for POL-SAR image segmentation, avoiding the choice of Gaussian kernel parameter and completing clustering effectively. From the experimental results, we can see that the proposed method has low time cost. Therefore, our method satisfied processing level for the land cover observation with use of SAR image. At the same time, our method not only keeps the overall classification accuracy, but also has more details of land cover. So our method can be applied to polarimetric SAR image recognition. However, there are still some problems like the edge blur. Our future work is to enhance the distinguish ability of the feature by adding other type features, such as texture and deep abstract features.

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China (Nos. 61472306, 91438201, 61572383), and the fund for Foreign Scholars in University Research and Teaching Programs (the 111 Project, No. B07048).

Conflicts of Interest

The authors declare no conflicts of interest.

[1] | Song, H., Yang, W., Xu, X., et al. (2014) Unsupervised PolSAR Imagery Classification Based on Jensen-Bregman LogDet Divergence. 10th European Conference on Synthetic Aperture Radar, 3-5 June 2014, Berlin, Germany. |

[2] |
Dabboor, M., Collins, M.J., Karathanassi, V., et al. (2013) An Unsupervised Classification Approach for Polarimetric SAR Data Based on the Chernoff Distance for Complex Wishart Distribution. IEEE Transactions on Geoscience and Remote Sensing, 51, 4200-4213. https://doi.org/10.1109/TGRS.2012.2227755 |

[3] |
Xu, Q., Chen, Q.H., Yang, S., et al. (2016) Superpixel-Based Classification Using K Distribution and Spatial Context for Polarimetric SAR Images. Remote Sensing, 8, 619. https://doi.org/10.3390/rs8080619 |

[4] |
Wang, P., Chen, W.S., Gou, S.P., Zhang, X.R., Li, X.F. and Jiao, L.C. (2017) A Weighted Joint Sparse of Three Channels Method for Full POLSAR Data Classification. IEEE International Geoscience and Remote Sensing Synposium (IGARSS), 23-28 July 2017, Fort Worth, 6146-6149.
https://doi.org/10.1109/IGARSS.2017.8128411 |

[5] |
Li, D.B., Gou, S.P. and Jiao, L.C. (2017) Full Polarization SAR Image Classification Using Deep Learning with Shallow Feature. IEEE International Geoscience and Remote Sensing Synposium (IGARSS), 23-28 July 2017, Fort Worth, 4566-4569.
https://doi.org/10.1109/IGARSS.2017.8128018 |

[6] |
Gou, S.P., Qiao, X., Zhang, X.R., Wang, W.F. and Du, F.F. (2014) Eigenvalue Analysis Based Approach for POL-SAR Image Classification. IEEE Transactions on Geoscience and Remote Sensing, 52, 805-818.
https://doi.org/10.1109/TGRS.2013.2244096 |

[7] |
Gou, S.P., Yang, X.F. and Li, X.F. (2016) Coastal Zone Classification with Full-Polarization SAR Imagery. IEEE Geoscience and Remote Sensing Letters, 13, 1616-1620. https://doi.org/10.1109/LGRS.2016.2597965 |

[8] |
Chen, W.S., Gou, S.P., Wang, X.L., Li, X.F. and Jiao, L.C. (2018) Classification of PolSAR Images Using Multilayer Autoencoders and a Self-Paced Learning Approach. Remote Sensing, 10, 110. https://doi.org/10.3390/rs10010110 |

[9] | Zhang, L.M., Zou, B., Zhang, J.P. and Zhang, Y. (2010) Classification of Polarimetric SAR Image Based on Support Vector Machine Using Multiple-Component Scattering Model and Texture Features. EURASIP Journal on Advances in Signal Processing. |

[10] | Kong, J.A., Swartz, A.A., Yueh, H.A., et al. (1988) Identification of Terrain Cover Using the Optimal Polarimetric Classifier. Journal of Electromagnetic Waves and Applications, 2, 171-194. |

[11] |
Cloude, S.R. and Pottier, E. (1997) An Entropy Based Classification Scheme for Land Applications of Polarimetric SAR. IEEE Transactions on Geoscience and Remote Sensing, 35, 68-78. https://doi.org/10.1109/36.551935 |

[12] | Pottier, E. and Cloude, S.R. (1995) Unsupervised Classification of Full Polarimetric SAR Data Using Target Decomposition Theorem and Entropy Analysis. Proceedings of International Geoscience and Remote Sensing Symposium, Florence, 10-14 July 1995, 2247-2249. |

[13] |
Cloude, S.R. (1985) Radar Target Decomposition Theorems. Electronics Letters, 21, 22-24. https://doi.org/10.1049/el:19850018 |

[14] |
Cloude, S.R. and Pottier, E. (1996) A Review of Target Decomposition Theorems in Radar Polarimetry. IEEE Transactions on Geoscience and Remote Sensing, 34, 498-518. https://doi.org/10.1109/36.485127 |

[15] | Cloude, S.R. and Pottier, E. (1997) Application of the H/A/á Polarimetric Decomposition Theorem for Land Classification. Proceedings of the SPIE, Wideband Interferometric Sensing and Imaging Polarimetry, San Diego, July 1997, 132-143. |

[16] | Pottier, E. (1996) The H/A/á Polarimetric Decomposition Approach Applied to PolSAR Data Processing. IEEE Transactions on Geoscience and Remote Sensing, 34, 498-518. |

[17] |
Freeman, A. and Durden, S. (1988) A Three-Component Scattering Model for Polarimetric SAR Data. IEEE Transactions on Geoscience and Remote Sensing, 36, 963-973. https://doi.org/10.1109/36.673687 |

[18] |
Lee, J.S., Grunes, M.R., Ainsworth, T.L., et al. (1999) Unsupervised Classification Using Polarimetric Decomposition and the Complex Wishart Classifier. IEEE Transactions on Geoscience and Remote Sensing, 37, 2249-2258.
https://doi.org/10.1109/36.789621 |

[19] | Pottier, E. and Lee, J.S. (1999) Application of the H/A/α Polarimetric Decomposition Theorem for Unsupervised Classification of Fully Polarimetric SAR Data Based on the Wishart Distribution. Proceedings of Committee on Earth Observing Satellites SAR Workshop, Toulouse, 26-29 October 1999, 335-340. |

[20] | Anfinsen, S.N., Jenssen, R. and Eltoft, T. (2007) Spectral Clustering of Polarimetric SAR Data with Wishart-Derived Distance Measures. Proceedings of the 3rd POLI NSAR Symposium, Frascati, 22-26 January 2007, SP-644. |

[21] |
Ersahin, K., Cumming, I.G. and Ward, R.K. (2010) Segmentation and Classification of Polarimetric SAR Data Using Spectral Graph Partitioning. IEEE Transactions on Geoscience and Remote Sensing, 48, 164-174.
https://doi.org/10.1109/TGRS.2009.2024303 |

[22] |
Gou, S.P., Qiao, X., Zhang, X.R., Wang, W.F. and Du, F.F. (2014) An Eigenvalue Analysis Based Approach for POL-SAR Image Classification. IEEE Transactions on Geoscience and Remote Sensing, 52, 805-818.
https://doi.org/10.1109/TGRS.2013.2244096 |

[23] | Carlos L.M. and Pottier, E. (2005) Study of the Speckle Noise Effects over the Eigen Decomposition of Polarimetric SAR Data. Proceedings of the 2nd International Workshop PoLinSAR, Fracati, 17-21 January 2005, 341-346. |

[24] |
Jin, Y.Q. and Cloude, S.R. (1994) Numerical Eigenanalysis of the Coherency Matrix for a Layer of Eandom Nonspherical Scatterers. IEEE Transactions on Geoscience and Remote Sensing, 32, 1179-1185. https://doi.org/10.1109/36.338366 |

[25] | Chen, T.G. and Li, Y.Z. (2005) Coherent Matrix Eigenvalue and Eigenvalue Index of Polarimetric SAR Images and Its Application of Bayes Classification. Information and Electronic Engineering, 3, 161-166. |

[26] |
Xiang, S.M., Nie, F.P. and Zhang, C.S. (2008) Learning a Mahalanobis Distance Metric for Data Clustering a Classification. Pattern Recognition, 41, 3600-3612.
https://doi.org/10.1016/j.patcog.2008.05.018 |

[27] |
Akbarizadeh, G. and Rahmani, M. (2015) A New Ensemble Clustering Method for PolSAR Image Segmentation. 7th Conference on Information and Knowledge Technology (IKT), Urmia, 26-28 May 2015, 1-4.
https://doi.org/10.1109/IKT.2015.7288775 |

[28] | Umut, O., Deniz, E. and Lan, T. (2006) Mean Shift Spectral Clustering for Perceptual Image Segmentation. IEEE International Conference on Acoustics, Speech, and Signal Processing, 2, 117-120. |

[29] | Strehl, A. and Ghosh, J. (2002) Cluster Ensembles—A Knowledge Reuse Framework for Combining Multiple Partitions. Journal on Machine Learning Research, 3, 583-617. |

[30] |
Fwlkes, C., Elongie, S. and Malik, J. (2001) Efficient Spatiotemporal Grouping Using the Nystrom Method. Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, 8-14 December 2001, 231-238. https://doi.org/10.1109/CVPR.2001.990481 |

[31] | Ouimet, M. and Bengio, Y. (2005) Greedy Spectral Embedding. Proceedings of 10th International Workshop on Artificial Intelligence and Statistics, Barbados, 6-8 January 2005, 253-260. |

[32] |
Zhang, K. and Kwok, J.T. (2009) Density-Weighted Nystrom Method for Computing Large Kernel Eigensystems. Neural Computation, 21, 121-146.
https://doi.org/10.1162/neco.2009.11-07-651 |

[33] | Drineas, P. and Mahoney, M.W. (2005) On the Nystrom Method for Approximating a Gram Matrix for Improved Kernel-Based Learning. Journal of Machine Learning Research, 6, 2153-2175. |

[34] |
Umut, O., Deniz, E. and Robert, J. (2008) Mean Shift Spectral Clustering. Pattern Recognition, 41, 1924-1938. https://doi.org/10.1016/j.patcog.2007.09.009 |

Journals Menu

Contact us

customer@scirp.org | |

+86 18163351462(WhatsApp) | |

1655362766 | |

Paper Publishing WeChat |

Copyright © 2023 by authors and Scientific Research Publishing Inc.

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.