A mutual information-Dempster-Shafer based decision ensemble system for land cover classification of hyperspectral data

Parham PAHLAVANI , Behnaz BIGDELI

Front. Earth Sci. ›› 2017, Vol. 11 ›› Issue (4) : 774 -783.

PDF (4124KB)
Front. Earth Sci. ›› 2017, Vol. 11 ›› Issue (4) : 774 -783. DOI: 10.1007/s11707-016-0611-2
RESEARCH ARTICLE
RESEARCH ARTICLE

A mutual information-Dempster-Shafer based decision ensemble system for land cover classification of hyperspectral data

Author information +
History +
PDF (4124KB)

Abstract

Hyperspectral images contain extremely rich spectral information that offer great potential to discriminate between various land cover classes. However, these images are usually composed of tens or hundreds of spectrally close bands, which result in high redundancy and great amount of computation time in hyperspectral classification. Furthermore, in the presence of mixed coverage pixels, crisp classifiers produced errors, omission and commission. This paper presents a mutual information-Dempster-Shafer system through an ensemble classification approach for classification of hyperspectral data. First, mutual information is applied to split data into a few independent partitions to overcome high dimensionality. Then, a fuzzy maximum likelihood classifies each band subset. Finally, Dempster-Shafer is applied to fuse the results of the fuzzy classifiers. In order to assess the proposed method, a crisp ensemble system based on a support vector machine as the crisp classifier and weighted majority voting as the crisp fusion method are applied on hyperspectral data. Furthermore, a dimension reduction system is utilized to assess the effectiveness of mutual information band splitting of the proposed method. The proposed methodology provides interesting conclusions on the effectiveness and potentiality of mutual information-Dempster-Shafer based classification of hyperspectral data.

Keywords

mutual information / Dempster-Shafer / hyperspectral / classification / support vector machine

Cite this article

Download citation ▾
Parham PAHLAVANI, Behnaz BIGDELI. A mutual information-Dempster-Shafer based decision ensemble system for land cover classification of hyperspectral data. Front. Earth Sci., 2017, 11(4): 774-783 DOI:10.1007/s11707-016-0611-2

登录浏览全文

4963

注册一个新账户 忘记密码

Introduction and background

The rich spectral information of hyperspectral images (HSI) increases the capability to distinguish different physical materials, leading to the potential of more accurate image classification. In this context, hyperspectral images have been successfully used for classification problems that require very precise description in spectral feature space. Extensive literature is available on the classification of HSIs. Maximum likelihood or Bayesian estimation methods (Jia, 2002), decision trees (Goel et al., 2003), neural networks (Del Frate et al., 2007), genetic algorithms (Vaiphasa, 2003), and kernel-based techniques (Müller et al., 2001; Camps-Valls and Bruzzone, 2005) have been widely investigated in this direction. One of the most popular classification methods is Support Vector Machines (SVMs) defined by Vapnik, which is a large margin based classifier with a good generalization capacity in the small-size training set problem with high-dimensional input space (Vapnik, 1998). Recently, SVMs have been successfully applied in the classification of hyperspectral remote-sensing data.Camps-Valls and Bruzzone (2005) demonstrated that SVMs perform equal or better than other classifiers in terms of accuracy on HSI.

Akbari proposed a spatial-spectral classification method based on marker based minimum spanning forest (MMSF) algorithm and SVM classification. The proposed method achieved improvement on different hyperspectral data (Akbari et al., 2016). Li developed a new framework for the classification of hyperspectral scenes that pursues the combination of multiple features. The ultimate goal of the proposed framework is to be capable to cope with linear and nonlinear class boundaries present in the data, thus following the two main mixing models are considered for hyperspectral data interpretation. In addition of the above crisp classification strategies, some research focuses on fuzzy classification of HSI (Li et al., 2015).

Each pixel from different satellites might represent several kilometres of land. In this sense, fuzzy classification would represent more natural mixtures and transition zones. Also, in the presence of mixed coverage pixels, crisp classifiers produced errors of omission and commission. Pepe and colleagues compared the performance of crisp and fuzzy classification on remotely sensed data. The results showed that fuzzy classifiers outperformed the crisp classifiers (Pepe et al., 2010). In our remote sensing field of interest, a number of approaches have been considered for fuzzy classification of hyperspectral data. Yu and colleagues applied a fuzzy k-nearest neighbour classification method for classification of HS data (Yu et al., 2002). Experiments on AVIRIS hyperspectral data in this paper confirmed higher effectiveness of fuzzy classifiers in comparison to that of crisp classifiers on HSI. Borasca applied a fuzzy version of SVM classifier on hyperspectral data (Borasca et al., 2006). In this research, the proposed approach outperformed crisp classification. Kasiri Bidhendi applied fuzzy unsupervised classification approaches on hyperspectral data. They used fuzzy C-means clustering (FCM) and fuzzy relational clustering (FRC). Zhang and Qiu successfully utilized an unsupervised neuro-fuzzy system for classification of hyperspectral data (Zhang and Qiu, 2012). They used Gaussian fuzzy self-organizing map (GFSOM). Di and colleagues applied fuzzy integral fusion on hyperspectral data in the aim of anomaly detection (Di et al., 2008).

A large number of features can become a curse in terms of accuracy if enough training samples are not available, i.e., due to the Hughes phenomenon in most of traditional classification techniques (Li et al., 2011). Conventional classification strategies often can-not overcome mentioned problem. Some researchers tried to use dimension reduction techniques to overcome this weakness. The existing dimension reduction methods mainly include two categories: the feature selection methods and the feature extraction methods. Imani and Ghasemian applied discriminant analysis and principal component analysis (PCA) as features extraction strategies on HSI (Imani and Ghasemian, 2015). The experimental results show good performance of their proposed method. Shen utilized a feature selection method based on particle swarm optimization (PSO) as feature selection method on HSI (Shen et al., 2015). Moreover, there exists significant research on dimension reduction techniques to overcome data redundancy of HSI. Nevertheless, the main drawback of feature selection methods is related to the loss of information through the elimination of some bands and the main drawback of feature extraction methods is related to change the spectral characteristics of HSI.

Recently, new research focuses on multiple classifier systems (MCS) to overcome the weaknesses of single classifiers (Benediktsson et al., 2007; Ceamanos et al., 2010). Ham successfully applied ensemble strategies to hyperspectral data (Ham et al., 2005). Also, Ceamanos et al. (2010) applied a multiple classifier system based on SVM for improvement of classification results on hyperspectral images. In their proposed method, the hyperspectral data set is decomposed into a few data sources according to the similarity of the spectral bands. Then, each source is processed separately by performing classification based on SVM. Finally, all outputs are used as input for final decision fusion performed by an additional SVM classifier. Results of the experiments show that the proposed multiple SVM system outperforms a standard classification of hyperspectral data.

This paper describes a new classifier ensemble approach based on mutual information (MI) and Dempster-Shafer (DS) to classify HSI. To address the capabilities of the proposed method, some conventional classification strategies are compared to the new results.

Proposed decision ensemble system based on mutual information and Dempster-Shafer

This paper proposes a decision ensemble system based on MI and DS for land cover classification of hyperspectral data. Figure 1 represents the general structure of the proposed methodology. First, MI is applied to group HSI cubes into partitions through a computation of the similarity measure of the spectral information. Second, the proposed methodology applies a fuzzy decision making system based on fuzzy maximum likelihood (FML) on each subset of hyperspectral bands. Then, a decision profile (DP) is established by using all fuzzy classifier decisions. Finally, a DS strategy is applied on DP to produce final classification result on HSI. In previous research, DS showed more accuracy and capability in comparison with those of other fusion procedures. The main motivations of the proposed methodology (especially in comparison with previous classification research) are considered as follows:

i) The proposed method tries to establish a decomposition of spectral information based on MI of spectral bands to overcome the high dimensionality problem of hyperspectral data. The main drawback of dimension reduction techniques in this field is related to the loss of information through the elimination of some bands in feature selection and to the changing of spectral information through transformation in feature extraction methods. By using MI, the proposed method tries to overcome these weaknesses by a system that enables the use of the entire high dimensional hyperspectral image space. Also, in experiments, we try to show this effectiveness by comparing the proposed method with conventional dimension reduction strategies.

ii) The proposed method applied a fuzzy decision making algorithm based on fuzzy maximum likelihood to overcome the presence of mixed coverage pixels in HSI. Also, the proposed method tries to compare this fuzzy system with SVM as one of the most powerful crisp kernel based classification methods.

iii) Because of the higher performance of the ensemble learning systems rather than single methods, the proposed method fused the ensemble of fuzzy classifiers based on Dempster-Shafer. In addition, this fuzzy ensemble system is compared with crisp ensemble of SVMs. In this crisp version, we utilized weighted majority voting (WMV) as fusion method on SVMs.

Spectral information splitting by mutual information

The basic principle of spectral information decomposition is that the spectral bands which have high similarity can be grouped into one group and the ones with little redundancy can be separated into different groups (Su et al., 2011). This paper applied MI as a similarity measure between spectral bands. It is known that independence between bands is one of the key issues to obtain relevant subsets of bands for classification purposes. The use of information measures, such as MI, in order to quantify the degree of independence provides a methodology to find generalized correlations among image bands. Not only is MI widely used as a criterion for measuring the degree of independence between random variables but it also measures how much a certain variable can explain the information content about another variable, i.e., being a generalized correlation measure. MI can be interpreted as a generalized correlation measure, which includes the linear and nonlinear dependence between variables (Chen et al., 2003).

Thus, this technique exploits this concept for band selection in order to reduce data redundancy and non-useful information. Martínez-Usó and Li applied MI as a successful similarity measure between spectral bands in feature selection (Martínez-Usó et al., 2006;Li et al., 2011). In addition, Bigdeli utilized MI measure as a band splitting strategy through a crisp classification of hyperspectral data (Bigdeli et al., 2013).

Entropy is a measure of uncertainty of random variables, and is a frequently used evaluation criterion of feature selection. If a discrete random variableX hasΦ alphabets and the probability density function is p(x),xΦ , the entropy of X is defined as:

H(X )= xΦ p(x)log p(x).

In the task of band grouping, the entropy of each band is computed by using all spectral information of this band. For two discrete random variablesX and Y, which have Φ and Ψ alphabets and their joint probability density function is p(x, y),x Φ, y Ψ, the joint entropy of X and Y is defined as:

H(X ,Y)= xΦ y Ψp(x, y)logp(x ,y) .

The MI is usually used to measure the correlation between two random variables and it is defined as

MI( X,Y) =H(X )+H(Y) H(X,Y) ,
where H(X ,Y) is the joint entropy.

The redundancy between two bands will be greater when the value of MI is larger. During the process of HSI cube splitting based on MI, the basic principle is that the bands are divided into groups according to local minima points of bands’ MI. These local minima points can be obtained automatically by comparing the neighbourhoods of every point.

Fuzzy maximum likelihood

The proposed method tries to benefit from a fuzzy classification strategy based on fuzzy maximum likelihood. FML is a fuzzy version of the conventional ML parameters: fuzzy mean and fuzzy variance-covariance matrix. The fuzzy mean can be defined as:

μc (x)=( i=1N f c( xi )x i )/( i=1N fc(xi )),

where N is the total number of patterns, fc is the membership function (MF) of class c, and x i is the i thpattern. The fuzzy variance-covariance matrix can be defined as

c = i= 1N f c( xi )( xi μ c) (xi μc )T i =1Nf c(xi ) .

The detailed description of FML was introduced in Chen (1999).

Dempster-Shafer as fuzzy decision fusion method

Decision ensemble systems are successfully applied on various types of data to improve single decision results. In the case of classification, these methods can improve classification accuracy in comparison to a single classifier by combining different classification algorithms or variants of the same classifier (Kuncheva, 2004). In such systems, a set of classifiers is produced at first and then combined by a specific fusion method. The resulting classifier is generally more accurate than any of the individual classifiers that make up the ensemble. The possible ways of combining the outputs of theL classifiers depend on what information (crisp or fuzzy) can be obtained from the individual members. In crisp category each classifier only outputs a unique class and finally a vector of classes is produced for each sample. Also the second type of classifier produces fuzzy output which means that in this case the classifier associates a confidence measurement for each class and finally produces a vector for every classifier and a matrix for the ensemble of classifiers. The output of the fuzzy classifiers that provides the class belongingness of an input pattern to different classes is arranged in a matrix form defined as a DP matrix. The DP matrix forL classifiers and C classes is shown in Fig. 2.

DS is one of the most powerful methods the can be applied on decision profile to fuse the results of fuzzy classifiers. Mathematical theory of evidence was first introduced by Dempster in the 1960’s, and later extended by Shafer (Shafer, 1976). This theory, which represents both imprecision and uncertainty, appears as a more flexible and general approach than Bayesian Theory. The following steps will summarize our DS as a fuzzy decision fusion method:

Step1: for j=1,...,C calculate the mean of the decision profile DP(x) of all members of wj from the data set Z. Call the mean a decision template DTj :

DTj = 1 Nj zk w j zk Z DP (Z k ),
where Nj is the number of element of Z from w j.

Step 2: after defining DT1,...,DTC , let DTji denote the ith row of decision template DTj . Denote by Di(x) the (soft label) output of Di, that is, Di (x)=[d i,1 (x),..., di,c( x)]T : the ith row of decision profile DP(x). We calculate the “proximity” Φ between D Tji and the output of classifier D i for the input x as follows:

Φ j,i(x) = (1+ DTjiDi(x ) 2)1 k=1C (1+ DTkiDi(x ) 2)1 ,
where . is any matrix norm. For example, we can use the Euclidean distance between the two vectors. Thus, for each decision template we haveL proximities.

Step 3: using previous equation we calculate for every class, j =1,...,c; and for every classifier, i=1 ,...,L , the following belief degrees:

bj( Di(x))= Φji(x ) kj (1Φ k,i (x)) 1 Φji (x)[1 kj (1Φ k,i (x)] .

Step 4: the final degree of support are

μj(x) =k i= 1L bj (Di(x)), j=1,...,c,
where K is a normalizing constant.

The most important reason for using DS is that this method is more powerful than other fuzzy decision fusion methods. Moreover, some previous research showed the high performance of this method (Kuncheva, 2004; Breve et al., 2007).

Study areas and data sets

The performance of the proposed methodology is tested on two well-known hyperspectral data sets. The first data is made up of a 145×145 pixel portion of the AVIRIS image acquired over north-western US, Indian Pine, acquired in June1992.

The Indian Pine data are available through Purdue University. The AVIRIS data set contains 220 spectral bands in the wavelength range 0.4–2.5mm although not all of the 220 original bands are employed in the experiments since 18 bands are affected strongly by atmosphere absorption phenomena and are consequently discarded. Hence, the dimensionality for the AVIRIS Indian Pine data set is 202. Figure 3 shows the original and the reference data of AVIRIS Indian Pine data. From the 16 different land cover classes available in the original ground truth of AVIRIS data; seven were discarded since only few training samples were available for them.

The remaining nine land cover classes were used to generate a set of training data and a set of testing data (Table 1).

The second data set is from Pavia University as another common hyperspectral data. This data set has been acquired by the German ROSIS sensor during a flight campaign over Pavia, northern Italy. Figure 4 represents the original and the reference data of ROSIS data. In ROSIS data, the number of spectral bands is 103 covering the wavelength range from 0.43mm to 0.86 mm. This data set contains 610×340 pixels with 1.3 m per pixel geometric sampling distance. Table 2 shows the reference data for Pavia University data. Training and testing samples were selected from different areas of the images. They are spatially disjointed.

Experimental results

To assess the performance of the proposed MI-DS system rather than dimension reduction, we use PCA and PSO, respectively, as common feature extraction and feature selection strategies. Also, a crisp decision ensemble system based on SVM as crisp classification and WMV as fusion method are utilized to compare with the proposed method. In all cases, we used the fixed training set to train classification methods.

Experiments with the AVIRIS Indian Pine data set

Based on the proposed method, it is necessary to split the hyperspectral data into subsets based on MI as similarity measure between spectral bands. Figure 5 shows the MI results.

Local minima points of this figure (black short line) correspond to the bands with low redundancy. Table 3 illustrates 12 subsets of spectral bands and their overall accuracies were obtained by FML and SVM classification methods.

By analysing Table 3 and Fig. 5, AVIRIS data was decomposed into 12 spectral information groups. Afterwards, FML and SVM were applied to classify each subset of HSI data. In Table 3, we reported overall accuracies for these two classification strategies. In more detail, FML represented higher accuracy for most of the spectral groups. However, for some of the groups, SVMs were slightly better than FML.

After classifying each subset separately, DS and WMV were applied to fuse the final decision of all subsets. In order to compare the proposed method with dimension reduction results, PCA and PSO were utilized on HSI spectral bands. Table 4 illustrates overall accuracies and kappa coefficients for mentioned classification strategies.

By comparing the different classification algorithms, it can be assessed that the proposed method based on MI for fuzzy version, i.e., DS, and crisp version, i.e., WMV, performed better than dimension reduction techniques. In particular, it is possible to observe that MI-DS system provided the best overall accuracy and kappa. The improvements of MI-DS are 7.4% and 8.5% for PCA and PSO respectively.

Figure 6 shows the final classification maps of MI-based decision ensemble systems and dimension reduction techniques. These classification maps confirm the results of Table 4. Table 5 gives the results in more detail for class by class accuracy.

As it can be seen, the proposed MI-DS based ensemble system is characterized by a better performance with respect to the other methods. First, better values of accuracies (overall accuracy, kappa and class accuracies) are obtained using the proposed MI-DS system. Then, for most of the classes PCA and PSO as dimension reduction strategies provided smaller accuracies. However, for “Soybean-no till” MI ensemble system based on weighted majority voting and for “Hay-windrowed” and “Woods” classes PCA provided better accuracy than the proposed MI-DS method.

Experiments with the ROSIS Pavia data set

The present paper tries evaluating classification accuracies of the proposed framework using ROSIS HSI data. Same as the previous experiment on AVIRIS data, Pavia hyperspectral data is converted into spectral subsets based on MI as similarity measure between spectral bands. Figure 7 illustrates the MI results.

As shown by Fig. 7, eight spectral subsets are obtained by the MI proposed method. In addition, Table 6 shows spectral subsets and overall accuracies obtained by fuzzy maximum likelihood and SVM on these spectral subsets.

Table 7 provides a comparison between the proposed MI-DS framework with dimension reduction systems and decision ensemble system based on WMV-MI.

The overall accuracies and kappa coefficients are similar to the previous experiment on AVIRIS data. The proposed fuzzy decision ensemble system based on the MI-DS system provided the best accuracy rather than the crisp decision ensemble system based on WMV and dimension reduction methods. This improvements are 11.2% and 14.04% for PCA and PSO, respectively. These results confirm that the proposed method based on the using of all spectral bands through an ensemble system based on MI-DS provides significant improvement rather than dimension reduction techniques. Figure 8 illustrates classification maps of the decision ensemble and dimension reduction methods.

Furthermore, Table 8 compared all classification strategies for class by class accuracy. These results exhibit the higher performance of the proposed MI-DS method for most of the classes. Further analysis of Table 8 shows that the mutual information based DS system produced higher accuracies for most of the classes; however, for “Bitumen” and “Shadows” classes dimension reduction based on PCA provided higher accuracies than the proposed method. This can be explained by smaller number of training samples for these two classes.

It should also be noted that the running time of the experiments has a major role in assessment of the results. First, because of the band splitting on hyperspectral data and then classification of each subset; i.e., in each single classification (12 times for AVIRIS and 8 times for ROSIS data), we did not use all spectral bands. Based on Tables 3 and 6, the number of spectral bands in each subset is small; consequently, classification of them takes just a few seconds. Regarding the above comment, the classification of subsets required only a few seconds, whereas the fusion of these classifiers needed more time. Second, comparison of the time of dimension reduction methods with decision fusion methods demonstrates that DS (as decision fusion method) method improved PSO 14% for ROSIS data. The time of the DS is 5 minutes longer than PSO, because PSO is a searching optimization method that requires more time. Also, PCA is a faster method than decision fusion methods and PSO. However, decision fusion methods provided 8%‒10% improvement over PCA in overall accuracy. Finally, it can be concluded that the proposed method provided more improvement in classification of hyperspectral data although it is slightly slower than PCA and PSO.

Discussion and conclusions

In this paper, we have developed a new framework for a fuzzy decision ensemble system on hyperspectral data which is based on MI and DS. A main contribution of the presented approach is the classification of HSI through a multiple decisions system without losing spectral information. In this work, we give a first step in computation of MI of spectral bands to split HSI into subsets. Then, we designed a fuzzy decision ensemble system based on maximum likelihood as the fuzzy classification technique and DS as the fuzzy decision fusion method. Also, the proposed method executed a crisp decision ensemble system based on SVM as a the crisp classification method and WMV as a crisp fusion method. In addition, two dimension reduction methods (PCA and PSO) were applied to compare with the proposed methodology. Our experimental results show some important points:

• Our band grouping system based on MI tries to solve the high dimensionality problem of HSI. Some previous research tried to select just useful bands in dimension reduction techniques to overcome data redundancies. Nevertheless, the main drawback of dimension reduction techniques is related to the loss of information through the elimination of some bands. The proposed method tries to overcome this weakness by a system that enables the use of the entire high dimensional hyperspectral image space. Comparison of the results in Tables 4 and 7 confirm the effectiveness on the proposed method over dimension reduction.

• Comparing the results of the crisp decision ensemble system based on SVM-WMV with fuzzy system based on DS show further improvement of our fuzzy proposed method. This is because of the existence of mixed coverage pixels in HSI that belong to more than one class.

The proposed method obtained approximately 97% and 98% overall accuracies for Indiana and Pavia hyperspectral data, respectively. Also, the proposed method improved common dimension reduction methods from 7% to 10%. These results suggest that the decomposition of the HSI classification problem into multiple decisions represents an effective way of improving the overall discrimination capability especially for fuzzy classification system.

References

[1]

Akbari DHomayuoni  SSaffari A, Mehrshad N (2016). Mapping urban land cover based on spatial-spectral classification of hyperspectral remote- sensing data. Int J Remote Sens37(2): 440–454

[2]

Benediktsson J A Chanussot J Fauvel M  (2007). Multiple classifier systems in remote sensing: from basics to recent developments. In: Haindl M, Kittler J, Roli F, eds. Multiple Classifier Systems. Heidelberg: Springer, 501–512

[3]

Bigdeli BSamadzadegan  FReinartz P  (2013). A multiple SVM system for classification of hyperspectral remote sensing data. Journal of the Indian Society of Remote Sensing41(4): 763–776

[4]

Borasca BBruzzone  LCarlin L Zusi M (2006). A fuzzy-input fuzzy-output SVM technique for classification of hyperspectral remote sensing images. Signal Processing Symposium, 2006. NORSIG 2006. Proceedings of the 7th Nordic, 2–5

[5]

Breve FPonti  MMascarenhas N  (2007). Multilayer Perceptron Classifier Combination for Identification of Materials on Noisy Soil Science Multispectral Images. Computer Graphics and Image Processing, 2007. SIBGRAPI 2007. XX Brazilian Symposium on. 239–244

[6]

Camps-Valls GBruzzone  L (2005). Kernel-based methods for hyperspectral image classification. IEEE Trans Geosci Rem Sens43(6): 1351–1362

[7]

Ceamanos XWaske  BBenediktsson J Chanussot J Fauvel M Sveinsson J  (2010). A classifier ensemble based on fusion of support vector machines for classifying hyperspectral data. International Journal of Image and Data Fusion1(4): 293–307

[8]

Chen C F (1999). Fuzzy training data for fuzzy supervised classification of remotely sensed images. Asian Conference Remote Sensing (ACRS 1999)

[9]

Chen HVarshney  P KArora  M K (2003). Performance of mutual information similarity measure for registration of multitemporal remote sensing images. IEEE Trans Geosci Rem Sens41(11): 2445–2454

[10]

Del Frate FPacifici  FSchiavon G Solimini C  (2007). Use of neural networks for automatic classification from high-resolution images. IEEE Trans Geosci Rem Sens45(4): 800–809

[11]

Di WPan  QHe L Cheng Y  (2008). Anomaly detection in hyperspectral imagery by fuzzy integral fusion of band-subsets. Photogramm Eng Remote Sensing74(2): 201–213

[12]

Goel P KPrasher  S OPatel  R MLandry  J ABonnell  R BViau  A A (2003). Classification of hyperspectral data by decision trees and artificial neural networks to identify weed stress and nitrogen status of corn. Comput Electron Agric39(2): 67–93

[13]

Ham JYangchi  ChenCrawford M M Ghosh J (2005). Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans Geosci Rem Sens43(3): 492–501

[14]

Imani MGhasemian  H (2015). Feature reduction of hyperspectral images: discriminant analysis and the first principal component. Journal of AI and Data Mining3(1): 1–9 

[15]

Jia X (2002). Simplified maximum likelihood classification for HS data in cluster space. In: IEEE Int. Geoscience and Remote Sensing Symp. (IGARSS ‘02), Toronto, Ontario, Canada5: 2578–2580

[16]

Kasiri bidhendi S Shirazi A Fotoohi N Ebadzadeh M  (2007). Material Classification of hyperspectral images using unsupervised fuzzy clustering methods. Third International IEEE Conference on Signal-Image Technologies and Internet-Based System, SITIS '07. 619 –623

[17]

Kuncheva L (2004). Combining Pattern Classifiers methods and algorithms. Hoboken, New jersey. Canada: John Wiley&sons, INC.

[18]

Li JHuang  XGamba P Bioucas-Dias J M B Zhang L Benediktsson J A Plaza A  (2015). Multiple feature learning for hyperspectral image classification. IEEE Trans Geosci Rem Sens53(3): 1592–1606

[19]

Li SWu  HWan D Zhu J (2011). An effective feature selection method for hyperspectral image classification based on genetic algorithm and support vector machine. Knowl Base Syst24(1): 40–48

[20]

Martínez-Usó A Pla FSotoca  J MGarcia-Sevilla  P (2006). Clustering based multispectral band selection using mutual information, Proceedings of the 18th International Conference on Pattern Recognition (ICPR)2: 760–763

[21]

Müller K L Mika SRätsch  GTsuda K Schölkopf B  (2001). An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw12(2): 181–201

[22]

 Pepe MBoschetti  LBrivio P A Rampini A  (2010). Comparing the performance of fuzzy and crisp classifiers on remotely sensed images: a case of snow classification. Int J Remote Sens31(23): 6189–6203

[23]

Shafer G (1976). A Mathematical Theory of Evidence. Princeton University Press

[24]

Shen JWang  CWang R Huang F Fan CXu  L (2015). A band selection method for hyperspectral image classification based on improved particle swarm optimization, International Journal of Signal Processing. Image Processing and Pattern Recognition8(4): 325–338

[25]

Su HYang  HDu Q Sheng Y  (2011). Semi-supervised band clustering for dimensionality reduction of hyperspectral imagery. IEEE Geosci Remote Sens Lett8(6): 1135–1139

[26]

Vaiphasa C (2003). Innovative genetic algorithm for hyperspectral image classification. In: Proc. Int. Conf. Map Asia.

[27]

Vapnik V N (1998). Statistical Learning Theory.New York: Wiley & Sons, Inc.

[28]

Yu SDe Backer  SScheunders P  (2002). Genetic feature selection combined with composite fuzzy nearest neighbor classifiers for hyperspectral satellite imagery. Pattern Recognit Lett23(1‒3): 183–190 

[29]

Zhang CQiu  F (2012). Hyperspectral image classification using an unsupervised neuro-fuzzy system. J Appl Remote Sens6(1): 1–15

RIGHTS & PERMISSIONS

Higher Education Press and Springer-Verlag Berlin Heidelberg

AI Summary AI Mindmap
PDF (4124KB)

1015

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/