Online MoG-LRMFmmp是什么意思思?

时刻准备越狱的猫喜欢的音乐 - 歌单 - 网易云音乐
时刻准备越狱的猫喜欢的音乐
播放:674次
网易云音乐多端下载
同步歌单,随时畅听320k好音乐
网易公司版权所有(C)杭州乐读科技有限公司运营:
违法和不良信息举报电话:6
举报邮箱:Robust Subspace Clustering via Penalized... (PDF Download Available)
See all >49 ReferencesSee all >2 Figures
2.83Xi'an Jiaotong University11.74Xi'an Jiaotong University+ 121.92Xi'an Jiaotong University42.51Xi'an Jiaotong UniversityShow more authorsAbstractMany problems in computer vision and pattern recognition can be posed as learning low-dimensional subspace structures from high-dimensional data. Subspace clustering represents a commonly utilized subspace learning strategy. The existing subspace clustering models mainly adopt a deterministic loss function to describe a certain noise type between an observed data matrix and its self-expressed form. However, the noises embedded in practical high-dimensional data are generally non-Gaussian and have much more complex structures. To address this issue, this paper proposes a robust subspace clustering model by embedding the Mixture of Gaussians (MoG) noise modeling strategy into the low-rank representation (LRR) subspace clustering model. The proposed MoG-LRR model is capitalized on its adapting to a wider range of noise distributions beyond current methods due to the universal approximation capability of MoG. Additionally, a penalized likelihood method is encoded into this model to facilitate selecting the number of mixture components automatically. A modified Expectation Maximization (EM) algorithm is also designed to infer the parameters involved in the proposed PMoG-LRR model. The superiority of our method is demonstrated by extensive experiments on face clustering and motion segmentation datasets.Discover the world's research15+ million members100+ million publications700k+ research projectsFigures
ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] Neurocomputing 0 0 0 (2017) 1–8 Contents lists available at ScienceDirect Neurocomputing journal homepage: www.elsevier.com/locate/neucom Robust subspace clustering via penalized mixture of Gaussians Jing Yao 1 , Xiangyong Cao 1 , Qian Zhao *, Deyu Meng, Zongben Xu School of Mathematics and Statistics and Ministry of Education Key Lab of Intelligent Networks and Network Security, Xi’an Jiaotong University, Xian 710049, PR China a r t i c l e i n f o Article history: Received 12 October 2016 Revised 5 April 2017 Accepted 21
May 2017 Available online xxx Keywo rds:
Subspace clustering Low-rank representation Mixture of Gaussians Expectation maximization a b s t r a c t Many problems in computer vision and pattern recognition can be posed as learning low-dimensional subspace structures from high-dimensional data. Subspace clustering represents a commonly utilized sub- space learning strategy. The existing subspace clustering models mainly adopt a deterministic loss func- tion to describe a certain noise type between an observed data matrix and its self-expressed form. How- ever, the noises embedded in practical high-dimensional data are generally non-Gaussian and have much more complex structures. To address this issue, this paper proposes a robust subspace clustering model by embedding the Mixture of Gaussians (MoG) noise modeling strategy into the low-rank representation (LRR) subspace clustering model. The proposed MoG-LRR model is capitalized on its adapting to a wider range of noise distributions beyond current methods due to the universal approximation capability of MoG. Additionally, a penalized likelihood method is encoded into this model to facilitate selecting the number of mixture components automatically. A modified Expectation Maximization (EM) algorithm is also designed to infer the parameters involved in the proposed PMoG-LRR model. The superiority of our method is demonstrated by extensive experiments on face clustering and motion segmentation datasets. (C)2017 Elsevier B.V. All rights reserved. 1. Introduction With dramatic development of techniques on data collection and feature extraction in recent years, the data obtained from real applications are often with high dimensionality. Such high- dimensional characteristic of data not only tends to bring large computation burden to the later data processing implementation, but also might possibly degenerates the performance of the uti- lized data process technique due to the curse of dimensionality is- sue. An effective strategy to alleviate this problem is to find the intrinsic low-dimensional subspace where such high-dimensional data intrinsically reside, and then make implementations on the low-dimensional projections of data. This is always feasible in real scenarios since features of practically collected data are generally with evident correlations. Such a “subspace learning” methodology has attracted much attention in the past decades and various re- lated methods have been proposed for different machine learning, computer vision and pattern recognition tasks [1–3] . *Corresponding author. E-mail addresses:
(J. Yao) ,
(X. Cao), timmy.zhaoqian@mail.xjtu.edu.cn , timmy. (Q. Zhao), dymeng@mail.xjtu.edu.cn (D. Meng), zbxu@mail.xjtu.edu.cn (Z. Xu). 1 1 indicates equal contribution In the recent years, a new trend on subspace learning, called subspace clustering, has appeared by simultaneously clustering data and extracting multiple subspaces, each corresponding to one data cluster [4–6] . Compared with traditional methods which as- sume data lie on a unique low-dimensional subspace, such “sub- space clustering” model better complies with many real scenar- ios where data are located on multiple subspace clusters. Typical applications include face clustering [7] , image segmentation [8] , metric learning [9] , feature grouping [10] and image representa- tion [11] . Accordingly this research has been attracting increasing attention in the recent years. Now let’s introduce the formal definition for the subspace clus- tering problem as below [4] : Definition 1.1 (Subspace Clustering) . Given a set of sampled data X = [ X 1 , . . . , X k ] = [ x 1 , x 2 , . . . , x n ] ∈ R d×n drawn from a union of k subspaces {S i } k i =1 , where X i denotes a collection of n i samples drawn from the subspace S i , and n = ? k i =1 n i . The task of subspace clustering is to cluster the samples according to the underlying subspaces they are drawn from. The main assumption underlying the subspace clustering is that each datum is sampled from one of several low-dimensional sub- spaces, and hence can be well represented as a linear combina- tion of the other data from the same subspace. The representa- tion matrix composed of all coefficients of such combinations is http://dx.doi.org/10.1016/j.neucom.2017.05.102 /(C) 2017 Elsevier B.V. All rights reserved. Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
2 J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] firstly taken as the encoding representation for the intrinsic sub- space clusters and then is utilized to extract subspace knowledge from data. This subspace learning task is mathematically formu- lated as follows: min C R (C ) + λL (X -~ X C ) , (1) where ~ X is a predefined dictionary, the first term regularize the representation matrix C to encode the subspace clustering prior knowledge in it, the second term is the loss function to fit the ad- ditive noise and λis a positive trade-off parameter. A natural choice for the loss function L (·) in Eq. (1) is the ? ·? F norm, which, from the probabilistic perspective, mainly character- ize Gaussian noise. However, real noises in applications are gen- erally non-Gaussian, and thus other types of loss functions were considered, such as ? ·? 1 norm and ? ·? 2, 1 norm 2 , correspond- ing to Laplacian noise and sample specific Gaussian noise, respec- tively [12] . These noise assumptions, however, still have limita- tions, since data noise in practical applications often exhibits much more complex statistical structures [13–16] . Therefore, a pre-fixed simple loss term is generally incapable of well fitting practical noise in data. The clustering accuracy of the utilized subspace clus- tering method [6] thus tends to be negatively influenced by this improper assumption. In this sense, it’s crucial to propose a robust subspace clustering model to tackle complex noise. To address this issue, this paper presents a novel subspace clus- tering method that is robust against a wider range of noise dis- tributions beyond traditional Gaussian, Laplacian or sample Gaus- sian noises. Our basic idea is to encode data noise as Mixture of Gaussians (MoG) in the subspace clustering model, and simulta- neously extract subspace cluster knowledge and adapt data noise. Here we prefer to employ MoG as the noise model due to its uni- versal approximation property to any continuous distribution [17] . Such idea is inspired by some recent noise modeling method- ology designed on some typical machine learning and computer vision tasks, such as low-rank matrix factorization (LRMF) [18] , robust principle component analysis (RPCA) [19] and tensor factor- ization [20] , and has been substantiated to be effective in the com- plicated noise scenarios. There exist multiple subspace clustering approaches recently, such as Sparse Subspace Clustering (SSC) [5] , Low-Rank Representative (LRR) [6] and Least Square Regression (LSR) [21] . In this work we readily adopt the LRR model [6,22] due to its utilization of clean data as the dictionary matrix, instead of taking original corrupted ones as most others, which better com- plies with the insight of subspace clustering. In addition, regard- ing the automatic selection of mixture Gaussian components, we propose a penalized MoG-LRR model through adopting a penalized likelihood technique inspired by [23,24] . We further design an ef- fective Expectation Maximization (EM) algorithm to infer all pa- rameters involved in this model. The superiority of the proposed method is substantiated on face clustering and motion segmenta- tion problems as compared with the current state-of-the-art meth- ods on subspace clustering. Specifically, the contribution of this work can be summarized as follows: o On subspace clustering, we integrated MoG noise modeling methodology into the LRR model, which enhances a robust sub- space clustering strategy with capability of adaptively fitting a wide range of data noises beyond current methods. o On MoG noise modeling methodology, through employing the penalized likelihood technique, the EM algorithm designed on the proposed MoG-LRR model is capable of automatically se- lecting a proper Gaussian mixture component number as well 2 The three norms are calculated as ? ·? F = ? N j=1 ( ? D i =1 (·) 2 ij ) , ? ·? 1 = ? N j=1 ( ? D i =1 | (·) ij | ) and ? ·? 2 , 1 = ? N j=1 ( ? D i =1 (·) 2 ij ) 1 2 , respectively. Tabl e 1 Some utilized notations in this paper. Notation Definition k Number of subspaces N Data size D Data dimensionality X = [ x 1 , . . . , x N ] Observed data A = [ a 1 , . . . , a N ] Clean data C = [ c 1 , . . . , c N ] Representation matrix E = [ e 1 , . . . , e N ] Sample-wise noise π= { π1 , . . . , πK } Mixture proportions ?={ ?1 , ... , ?K } Covariance matrices as other involved parameters in this model. This also prompts the frontier of noise modeling and makes it easy for the selec- tion of this important parameter. Our paper is organized as follows. Related work is introduced in Section 2 . Section 3 proposes our model called Penalized Mixture of Gaussians Low-Rank Representation (PMoG-LRR) and then presents a modified EM algorithm for solving this model. Section 4 presents experimental results implemented on synthetic and real data sets to substantiate the superiority of our proposed method over other state-of-the-arts. Finally, a brief conclusion is drawn in Section 5 . Throughout the paper, we denote scalars, vectors, matrices as the non-bold, bold lower case and bold upper case letters, respectively. Some notations used in this paper are summarized in Table 1 . 2. Related work The past two decades has witnessed a rapid development in the field of subspace clustering. The related methods can be roughly classified into four categories: algebraic methods, iterative meth- ods, statistical methods, and spectral-clustering-based methods [4] . Algebraic methods, typically represented by Matrix Factorization-based methods [25–27] , first find a permutation matrix and calculate the multiplication matrix of data and the permutation matrix, and then factorize this multiplication matrix into two rank- r matrices, a base-matrix and a block diagonal matrix, respectively. But these methods are generally sensitive to noise and require the knowledge of the rank r of data matrix. The iterative methods, e.g., K-subspaces [28] use an iterative way to model and segment data. Specifically, such methods first assign data to pre-defined multiple subspaces, and then update the subspaces and reassign each data point to the closest subspace. The drawback of above two methods is that they incline to be sensitive to initialization and outliers. Besides, they need to know the number of subspace and their corresponding dimensions in advance. The statistical methods, e.g., Mixtures of Probabilistic PCA (MPPCA) [29] , assume that the sample data are generated from a Mixture of Gaussians (MoG) distribution and then uses the Expectation Maximization (EM) algorithm to update the data segmentation and model parameters alternatively under the Maxi- mum Likelihood Estimation (MLE) framework. One disadvantage of these methods is that the model always cannot fit the cases that the intrinsic distributions of the data inside each subspace are not Gaussian. Recently, spectral-clustering-based subspace clustering methods has been attracting more attention [5,6,21] due to its rational methodology and successful performance in applications [30] . The fundament of these methods is to assume that each data point can be linearly represented by all the other data points from the same subspace cluster. These methods generally contain two steps. Firstly, an affinity matrix is built to capture the similarity between pairs of data points. Then, the segmentation of data is obtained by applying spectral clustering algorithm [31] to the affinity matrix. Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 3 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] Since this methodology is the main focus of this work, we give a more detailed introduction to its recent developments as follows. Elhamifar and Vidal [5] proposed the Sparse Subspace Cluster- ing (SSC) to find the sparsest representation of each data point by utilizing the l 1 norm regularization term. Low-Rank Representation (LRR) [6] aims to get a low rank representation for robust subspace recovery of the data containing corruptions by imposing the nu- clear norm on the representative matrix. Besides, based on the l 2 norm, Least Squares Regression (LSR) [21] was proposed for sub- space clustering under Enforced Block Diagonal (EBD) conditions. Low-rank subspace clustering (LRSC) [22] extends both methods and proposes an alternative non-convex formulation, which can be solved efficiently by a closed-form solution when there is no out- liers. To improve clustering accuracy, multiple amelioration tech- niques on subspace learning, including Bayesian method [32] , quadratic programming [33] , manifold regularization [12] and Markov random walks [34] , have also been recently attempted to further ameliorate the capability of subspace clustering. Here, we list some typical instances as follows: Based on block-diagonal structure prior on ideal representation matrix, Feng et al. [35] at- tempt to explicitly pursue such a block diagonal structure by proposing a graph Laplacian constraint based formulation. In ad- dition, to handle more complex noise, many robust subspace clus- tering methods have been studied recently. Both Lu et.al. [36] and He et.al. [37] use correntropy as a robust measure to suppress the influence of large corruptions and make subspace clustering model be robust to non-Gaussian noises. The MoG noise modeling strategy was firstly considered in LSR subspace clustering model in [13] . However, this strategy is inferior in that it needs to empir- ically while not automatically tune the number of mixture Gaus- sian components, and besides, in the method the representation matrix between data is constructed on the corrupted data while not clean ones, which tends to make it still sensitive in the pres- ence of noises or outliers, as substantiated by our experiments in Section 4 . 3. Robust subspace clustering by penalized mixture of Gaussians In this section, we first briefly introduce the LRR subspace clus- tering model [6] . Then we propose our PMoG-LRR model. Finally, we design an efficient EM algorithm to solve this model. 3.1. LRR model revisit To better illustrate the main idea of our method in the context of LRR, we briefly review the basic idea of LRR in this subsection. Let’s consider the case that data X are drawn from a union of multiple subspaces given by ∪ k i =1 S i , where S 1 , S 2 , . . . , S k represent the low-dimensional subspaces of underlying data. The traditional LRR model can be formulated as the following rank minimization problem: min C , E rank (C ) + λ|| E || p s.t X = XC + E , (2) where the constraint means a sample can be linearly represented by the ones located in the same cluster with it, E indicates the noises embedded in data, || ·|| p denotes a certain loss function, like L 2 -norm, L 1 -norm and L 2, 1 -norm losses, and the parameter λ& 0 is the compromising parameter between two objective terms. Un- fortunately, (2) is both non-smooth and non-convex, and difficult to solve. Generally, a tractable optimization problem is utilized by replacing the rank term with the nuclear norm || C || *= ? i σi (C ) , where σi ( ·) denotes the i th singular value of a certain matrix. Then the first term of the LRR model turns to be a convex com- ponent as follows: min C , E || C || *+ λ|| E || p s.t X = XC + E . (3) Note that in traditional LRR as well as most other subspace clustering models, the observed data X is directly used as the basis matrix in the model. This is actually improper since the expected basis matrix should be the clean matrix without noises underly- ing X . An ameliorated version of LRR is thus recently proposed in [22] to deal with this issue. The modified LRR model is min A , C , E || C || *+ λ|| E || p s.t A = AC , X = A + E , (4) where A represents the clean data matrix underlying X . It is evi- dent that the new model better formulates the subspace represen- tation correlation, and we thus prefer to employ this model as the baseline in our work. 3.2. Formulation of PMoG-LRR model In this subsection, we will introduce our PMoG-LRR model. We utilize the MoG distribution to model the real noise, which con- structs a universal approximator to any continuous density func- tion in theory [17] . Specifically, we assume that each column e n (n = 1 , 2 , . . . , N) of E in (3) follows an MoG, i.e., P (e n ) = K ? k =1 πk N (e n ;0 , ?k ) , (5) where πk is the mixing proportion with πk ≥0 and ? K k =1 πk = 1 , K is the number of the mixture components and N (e n | 0 , ?k ) de- notes the k th Gaussian component with zero-mean and covariance matrix ?k . To reduce the possible over-fitting issue, each covari- ance matrix ?k is assumed to be a diagonal matrix and is rep- resented as ?k = diag(σ2 kj ) , j = 1 , 2 , . . . , D . In addition, we denote π= (π1 , π2 , . . . , πK ) ? and ?= (?1 , ?2 , . . . , ?K ) and assume each column of E is independent and identically distributed (i.i.d). Then, the likelihood function can be written as P (E ;π, ?) = N ? n =1 K ? k =1 πk N (e n ;0 , ?k ) , (6) and the negative log-likelihood function is -log P (E ;π, ?) = -N ? n =1 log ? K ? k =1 πk N (e n ;0 , ?k ) ? . (7) As aforementioned, inspired by [22] , given the corrupted data matrix X , instead of utilizing this matrix as the basis matrix, we directly employ the clean data A underlying X , i.e., A = AC . The re- lationship between two matrices is: X = A + E . In our model, we further consider the crucial issue of how to automatically select the number of mixture components K . Vari- ous model selection techniques can be readily employed to resolve this issue. Most conventional methods are based on the likelihood function and some information theoretic criteria, such as AIC and BIC. Here we adopt a penalized likelihood method [24] , which has been proved to be empirically effective and theoretically sound, to shrink the mixing weights continuously and retain effective com- ponents in the mixture distributions. This penalty function is de- fined as Q ( π;β) = nβK ? k =1 p k [ log (?+ πk ) -log (?)] , (8) where n is the number of columns of the data matrix, namely the number of the data, ?is a very small positive number, βis a tun- ing parameter ( β≥0), and p k is the number of free parameters for the k th component. Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
4 J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] We can then formulate the PMoG-LRR model as: min A , C , E , π, ?|| C || *-λ2 [ log P (E ;π, ?) -Q ( π;β)] s.t A = AC , X = A + E , (9) πk ≥0 , K ? k =1 πk = 1 , ?k = diag(σ2 kj ) ∈ S + , k = 1 , . . . , K, where S + represents the set of positive definite matrices. 3.3. Modified EM Algorithm for solving PMoG-LRR model An elegant and powerful way to solve (9) is the Expectation Maximization (EM) algorithm, which aims to find the maximum- likelihood estimation of the parameters iteratively. In this subsec- tion, we develop a modified EM algortihm for solving the proposed model. In the E step, we compute the responsibility parameters γnk based on the current parameters ?(t) = { X (t) , A (t) , K (t) , π(t) , ?(t) } : γnk = πk N (e n ;0 , ?k ) ? K l=1 πl N (e n ;0 , ?l ) , (10) where e n = x n -a n , x n and a n denote the n th columns of X and A , respectively, and the superscript denotes the iteration number. Based on the responsibility so obtained, we can compute the Q- function Q (?, ?(t) ) as Q (?, ?(t) ) = N ? n =1 K ? k =1 γnk [ log N (e n ;0 , ?k ) + log πk ] -nβK ? k =1 p k [ log (?+ πk ) -log (?)] . (11) In the M step, we need to maximize (11) to achieve the updated MoG parameters ?(t+1) . We first cope with problem of updating ?k , k = 1 , 2 , . . . , Kby optimizing the following problem: min ?k -N ? n =1 K ? k =1 γnk log N (e n ;0 , ?k ) s.t ?k = diag(σ2 kj ) ∈ S + , j = 1 , 2 , . . . , D. (12) By taking the first derivative of the objective function with respect to ?k , we obtain ?(t+1) k = d iag ? d iag ? 1 n k ? N ? n =1 γnk e n e T n + ηI D ×D ? ? ? , (13) where n k = ? N n =1 γnk , η& 0 is a small regularization parameter to make ?(t+1) k invertible, and I D ×D is an identity matrix with D di- mensions. To update πk , k = 1 , . . . , K, we need to introduce a Lagrange multiplier αto take the constraint ? K k =1 πk = 1 into consideration, and then minimize -N ? n =1 K ? k =1 γnk log πk + α? ? k πk -1 ? -nβD K ? k =1 [ log (?+ πk ) -log (?)] . (14) Setting the derivative of (14) with respect to πk to zero and con- sidering that ?is close to zero, we obtain: π(t+1) k = max ? 0 , 1 1 -βKD ? 1 n N ? n =1 γnk -βD ? ? . (15) To update A and C , we should solve the following sub- problem: min A , C || C || *-λ2 N ? n =1 K ? k =1 γnk log N (x n -a n ;0 , ?k ) s.t A = AC , (16) which can be mathematically reformulated as: min C , ~ A || C || *+ λ2 || ~ X -~ A || 2 F s.t ~ A = ~ A C , (17) where ~ X = [ ~ x 1 , ~ x 2 , . . . , ~ x N ] , ~ x n = H 1 2 n x n , ~ A = [ ~ a 1 , ~ a 2 , . . . , ~ a N ] , ~ a n = H 1 2 n a n and H n = ? K k =1 γnk ?-1 k . From (17) , we can obtain a closed-form solution by applying Lemma 2 and 3 in [38] . Specifically, the solution to (17) is ~ A = U r S r V T r , C (t+1) = V r V T r , (18) where S r represents the top r = arg min k (k + λ2 ? i&k σ2 i ) singular values, U r and V r are the matrices composed by the correspond- ing left and right singular vectors of ~ X , respectively. It should be noted that here in the limiting case, r can also be equivalently de- termined by thresholding the singular value σi at ? 2 /λ. Then, we can update A by a (t+1) n = H -1 2 n ~ a n , n = 1 , 2 , . . . , N. (19) The procedure of the proposed EM algorithm is summarized in Algorithm 1 . Algorithm 1 EM Algorithm for PMoG-LRR. Input: Data matrix X , initial model parameters ?(0) ={mixture param- eters π(0) and ?(0) ,initial component number K (0) , clean data matrix A (0) , representation matrix C (0) } and t = 1 . Output: The representation matrix C and the clean data matrix A . 1: repeat 2: ( E step ): Update γ(t) via (10). 3: ( M step for ?): Update ?(t) via(13). 4: ( M step for π): Update π(t) via(15). 5: ( M step for C , A ): Update C (t) and A (t) via (18) and (19), re- spectively. 6: t ← t + 1 . 7: until converge ; 3.4. Implementation details In our experiments, the compromising parameter λbetween regularization term and penalized likelihood term is determined by cross validation. Additionally, we provide a series of tuning pa- rameter βwith alternative values. Then the Bayesian Information Criterion (BIC) or the Schwarz criterion [39] is used to select the βwith largest BIC value. We set initialized mixture components K (0) as 5 throughout our experiments. For each component, we sam- ple from a standard normal distribution to initialize the diagonal elements of covariance matrix ?(0) . We use U 0 D 0 V 0 to initialize A (0) , where D 0 is the non-zero part of singular values of X , U 0 and V 0 are matrices composed by the corresponding left and right sin- gular vectors, respectively. Finally, based on the theory in [26,40] , suppose the skinny SVD decomposition of A (0) is U ?V , and then C (0) is initialized with VV ? . Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 5 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] Fig. 1. The face images of the fifth subject from the AR database. 3.5. Convergence We proposed an EM type algorithm for solving optimization problem (9) . During the iterations, each involved subproblem is convex and has global solution. Therefore, by virtue of the gen- eral theory of EM algorithm [41] , we can conclude that the ob- jective function of (9) monotonically decreases. Consequently, the proposed algorithm converge in the sense of objective value. 4. Experimental results To evaluate the performance of our proposed PMoG-LRR method, we conducted a series of experiments on two applications of subspace clustering: face clustering and motion segmentation. Seven subspace clustering methods were considered for compari- son, including local subspace affinity (LSA) [42] , SSC [5] , LRR [6] , LSR [21] , block-diagonal LRR (BD-LRR) [35] , Low-Rank Subspace Clustering (LRSC) [22] and MoGR [13] . These methods represent the current state-of-the-art along this research line. We utilize two existing metrics for quantitative evaluation. Let l i and g i be the clustered label and ground truth label of the i th sample, respectively, and then clustering accuracy [5] is defined as Accuracy = ? N i =1 1 (g i ,map(l i )) N ×100 , where map ( l i ) denotes the best map from obtained label l i to equivalent label in ground truth and 1 ( a , b ) is the indicator function that equals to one if a = band zero otherwise. For two clusters C and C ? , representing ground truth and calcu- lated clustering results, respectively, we can compute their mutual information (MI) as: MI(C, C ? ) = ? c i ∈ C,c ? j ∈ C ? p(c i , c ? j ) log 2 p(c i ,c ? j ) p(c i ) p(c ? j ) , where p ( c i ) and p(c ? j ) denote the probabilities that an arbitrary data point belongs to the clusters c i and c ? j , respectively, and p(c i , c ? j ) denotes the joint probability that an arbitrary data point belongs to the clusters c i and c ? j at the same time. MI ( C , C ? ) can be used to determine how similar the clusters C and C ? are. It takes values between zero and max ( H ( C ), H ( C ? )), where H ( ·) de- notes the entropy. So we normalize the mutual information to simplify the comparison between different pairs of cluster sets as NMI(C, C ? ) = MI(C,C ? ) max (H(C) ,H(C ? )) , which equals to one when two sets of clusters are identical and zero when the the two clusters have not shared even one data point. All parameters involved in the competing methods were empir- ically tuned or specified as suggested by the related literatures to guarantee their possibly good performance. Some experimental re- sults of existing methods reported in related papers have been di- rectly utilized. All experiments are carried out on a Windows sys- tem with 3.6GHz Intel Core i7 processor using MATLAB2014b. 4.1. Face clustering Face clustering is to cluster face images collected from multi- ple individuals based on their identities. The AR Face Database 3 3 http://www-sipl.technion.ac.il/new/DataBases/Aleix%20Face%20Database.htm . [43] and Extended Yale Dataset B 4 [44] are adopted as the bench- mark datasets for this task. 4.1.1. AR face dataset The AR Face Database contains over 4,0 0 0 facial images from 126 subjects (70 men and 56 women). For each subject, 26 facial images are taken in two separate parts for training and testing, re- spectively. These images suffer different facial variations, including various facial expressions (neutral, smile, anger, and scream), illu- mination variations (left light on, right light on, and all side lights on), and occlusion by sunglasses or scarf. Fig. 1 shows some typical samples from the dataset for demonstration. We select images of first c = { 5 , 8 , 10 } subjects and apply all competing methods to implementing them. We also adopt dimen- sion reduction procedure with standard PCA in this scenario. The clustering results including accuracy and NMI are shown in Table 2 . Fig. 2 demonstrates the corresponding affinity matrix constructed from different algorithms. From Table 2 , it can be easily seen that our method performs better than other competing methods in both the clustering ac- curacy and NMI, and takes evidently less time than most of the competing methods, while only unsubstantially slower than SSC, LSR and LRSC. This shows that our method can perform robust in presence of noise or outliers. The superiority of the proposed method can also be verified visually by observing Fig. 2 . It is easy to see that the affinity matrices obtained by LRR, BD-LRR, LRSC, MoGR and our method have more evidently depicted the config- uration of subspace clusters as compared with those obtained by LSA and SSC, which lack of within-cluster connectivities. It can also be observed that the contrast between within-cluster and inter- cluster relationships have been significantly magnified by MoGR and PMoG-LLR, which is attributed to the better modeling of noise. Besides, those clusters are not perfectly detected by MoGR, e.g., cluster 2 and 6, have been more clearly recognized by PMoG-LRR, with evidently larger representation coefficients. This better visual result, as compared with those of other methods, indicates that our method has better ability of clustering the correlated data located in the similar subspace. 4.1.2. Extended Yal e dataset The Extended Yale
Database B consists of 2,414 frontal face im- ages of 38 subjects, where there are 64 faces for each subject, ac- quired under different lighting, poses, and illumination conditions. Each image is cropped into 192 ×168
pixels. Fig. 3 shows some typical samples from this dataset. Following the experimental setting in [5,35] , we build four types of tasks according to the following scheme. First, 38 sub- jects are separated into four groups: subjects 1 to 10, subjects 11
to 20, subjects 21 to 30, and subjects 31
to 38. Then we consider choice of c = { 2 , 5 , 8 , 10
} subjects for each of the first three groups and c = { 2 , 5 , 8 } for the last group. We
implement all the subspace clustering algorithm on each set of c subjects. To reduce the computational cost, each face image is resized to a resolution of 48 ×42 pixels. As shown in [45] , the faces with 4 http://vision.ucsd.edu/leekc/ExtYaleDatabase/ExtYaleB.html . Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
6 J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] Tabl e 2 The average clustering accuracy (%), NMI (%) and running time (s) on the AR database for all competing methods. Method LSA SSC LRR LSR BD-LRR LRSC MoGR PMoG-LRR 5 Subjects Accuracy 73.77 83.08 84.23 93.85 86.40 93.08 93.85 95.38 NMI 74. 68 82.17 82.97 86.79 83.92 83.60 91.11 92.93 Time 2.22 0.89 2.24 1.3 3
5.99 1.2 1
8 Subjects Accuracy 66.78 75.00 79.81 80.77 83.21 90.38 90.38 94.23 NMI 69.79 75.54 78.37 83.70 80.38 85.78 88.64 92.22 Time 5.62 1.05 4.54 1.6 7 14 .17 1. 5 0
34.70 1.9 2
10 Subjects Accuracy 68.31 77.69 84.23 79.23 85.68 92.69 88.85 93.46 NMI 72.47 78.74 83.38 81.15 85.31 89.81 87.96 91.11 Time 8.90 1.25 6.79 2.19 22.16 2.07 111.13
4.67 --Fig. 2. The affinity matrices of 10
obtained by all competing methods from the AR database. Fig. 3. Some face samples of one subject from the YALE B database. different types of diffusion or background illumination fall into a 4-dimensional subspace. So before the vectorized image data are inputted into our algorithm, we also resort to standard PCA and project our original vectorized data onto 4 ×c dimensional sub- space. In order to make a fair comparison, the post-processing pro- cedure is not performed in this scenario. The detailed results are presented in Table 3 . From Table 3 , it is easy to observe that PMoG-LRR achieves the best clustering accuracy in almost all cases except the 2 subject case, where it performs the second best, slightly worse than BD- LRR in accuracy Mean and Median. Besides, compared with the tra- ditional LRR, our method obtains at least 5% increase of the clus- tering accuracy. This substantiates the superiority of the proposed method in such face modeling scenarios. 4.2. Motion segmentation Motion segmentation aims to segment the trajectories as- sociated with n different moving objects into different groups according to their motions in a video sequence. The Hopkins 155
dataset 5 [46] is one of the most commonly utilized bench- mark databases to evaluate performance of subspace clustering algorithms. It consists of 155 video sequences, where 120 ones contain 2 motions and 35 ones have 3 motions. For each se- quence, a tracker is used to extract feature trajectories and out- liers are removed manually. Some examples with extracted fea- tures are shown in Fig. 4 . As traditional pre-processing procedures for this task, we adopt PCA to project the original data into a 12-dimensional subspace according to [5] . The procedure of sub- space clustering is performed on the trajectory spatial coordinates of each sequence. The results of the clustering results using differ- ent methods are shown in Table 4 . From this table, we can observe that PMoG-LRR achieves the highest clustering accuracy in most cases except slightly worse than LSR and BD-LRR in terms of accuracy Mean. For all 155 video sequences, our method achieves better clustering accuracy in Min, Median and Std., except the second best in Mean. Note that al- though MoGR has also considered MoG noise modeling strategy in LSR model, it still cannot perform as robust as the proposed method. This is possibly because they compute the coefficient ma- trix on the basis of the original corrupted data while not the clean one, which possibly degenerates its performance. Also we should note that compared with the baseline LRR method, our method achieves an evident gain in clustering accuracy (about 3.5%) for all 155
video sequences, indicating that MoG noise modeling signifi- cantly enhances the capability of LRR to make it perform robust to complex scenarios and fit wider range of noises. 5 http://www.vision.jhu.edu/data/hopkins155/ . Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 7 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] Tabl e 3 The clustering accuracies (%) on the Extended Yale
Dataset B for all competing methods. Method LSA SSC LRR LSR BD-LRR LRSC MoGR PMoG-LRR 2 Subjects Min 64.09 86.14 85.63 76.73 90.01 86.08 79.48 90.17 Mean 68.92 89.43 88.14 79.66 93.19 90.43 84.71 93.10 Median 70.04 90.10 88.91 80.63 92.48 90.69 85.09 92.24 5 Subjects Min 52.18 74.89
80.77 72.08 82.08 79.31 72.9 3 85.17 Mean 55.67 81.59 85.01 76.91
86.03 84.76 78.24 90.17 Median 54.93 82.19 85.13 76.17 85.73 85.69 80.34 87.93 8 Subjects Min 39.87 64.11 62.30 68.09 70.94 64.08 67.71 74.35 Mean 42.05 68.09 68.76 71.0 3 73.0 0 70. 01 70.38 76.29 Median 43.11 68.62 69.32 70.14 73.34 70.4 9 69.56 76.08 10 Subjects Min 35.87 49.31 57.06
56.14 62.08 60.15 61.66 63.24 Mean 37.69 59.08 61.28 61.0 4 64.49 62.47 63.08 68.66 Median 37.01 62.96 60.37 59.89 64.77 62.01 63.85 70.00 Fig. 4. Some samples with the extracted features from the Hopkins 155 dataset. Tabl e 4 The clustering accuracies (%) on the Hopkins 155 database for all competing methods. Method LSA SSC LRR LSR BD-LRR LRSC MoGR PMoG-LRR 2 Motions Min 59.10 51.10 59.70 63.64 75.07 69.47 63.40 78.36 Mean 93.20 96.30 96.80 99.60 99.31 91.36 96.76 98.43 Median 97.20 10 0.0 0 99.70 10 0.0 0 10 0.0 0 92.43 98.34 10 0.0 0 Std. 8.00 9.70 8.20 6.80 6.47 10. 45 7.16
6.02 3 Motions Min 53.40 55.40 58.50 65.44 73.39
65.39 60.51 75.28 Mean 83.20 88.60 92.20 93.37 96.31 87.84 95.03 96.55 Median 84.40 96.70 97.20 98.79 99.20 89.55 97.51 99.17 Std. 12.60
15. 0 0 10 .30
9.97 7.0 1 15. 44
6.81 All Min 53.40 51.10 58.50 63.64 73.3 9 65.39 60.51 77.36 Mean 90.94 94.56 95.76 98.19 98.63 90.57 96.40 98.23 Median 95.20 10 0.0 0 99.33 99.69 10 0.0 0 90.47 98.47 10 0.0 0 Std. 10 .10 11. 6 0
8.90 8.62 6.56 12.73
10. 31 5.96 5. Conclusion In this paper, we propose a robust subspace clustering method by utilizing the MoG noise modeling strategy and the self- expressed property of clean data basis. Besides, we also adopt a penalized likelihood method to learn the mixture component au- tomatically. Compared with the current subspace clustering meth- ods, our method performs more robust and achieves the state-of- the-art subspace clustering performance in the real complex noise scenarios, including face clustering and motion segmentation. In future research, we will try to to extend the noise model- ing methodology to more machine learning and pattern recogni- tion tasks, e.g., tensor subspace clustering [47] , and will investigate more theoretical insights under such subspace clustering frame- work. References [1] D. Meng , Q. Zhao , Z. Xu , Improve robustness of sparse PCA by L1-norm maxi- mization., Pattern Recognit. 45 (1) (2012) 4 87–4 97 . [2] F. De La Tor re
, M.J. Black , A framework for robust subspace learning, Int. J. Comput. Vision 54 (1–3) (2003) 117–142 . [3] Y. Li , On incremental and robust subspace learning, Pattern Recognit. 37 (7) (2004) 1509–1518 . [4] R. Vidal , A tutorial on subspace clustering, IEEE Signal Process. Mag. 28 (2) (2010) 52–68 . [5] E. Elhamifar , R. Vidal , Sparse subspace clustering: algorithm, theory, and applications, IEEE Trans. Pattern Anal. Mach. Intell. 35 (11) (2013) 2765–2781 . [6] G. Liu , Z. Lin , S. Yan , J. Sun , Y.
Yu , Y. Ma , Robust recovery of subspace struc- tures by low-rank representation, IEEE Trans. Pattern Anal. Mach. Intell. 35 (1) (2013) 171–184 . [7] J. Ho , M.-H. Yang , J. Lim , K.-C. Lee , D. Kriegman ,Clustering appearances of ob- jects under varying illumination conditions, in: Proceedings of the IEEE Con- ference on Computer Vision and Patter n Recognition, 2003, I - 11-18 . Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
8 J. Yao et al. / Neurocomputing 0 0 0 (2017) 1–8 ARTICLE IN PRESS JID: NEUCOM [m5G; September 15,
] [8] A.Y. Yan g , J. Wri ght , Y.
Ma , S.S. Sastry , Unsupervised segmentation of natu- ral images via lossy data compression, Comput. Vision Image Underst. 110 (2) (2008) 212–225 . [9] J. Wa ng , Z. Deng , K.-S. Choi , Y. Jiang , X. Luo , F. -L . Chung , S. Wang , Distance met- ric learning for soft subspace clustering in composite kernel, Pattern Recognit. 52 (1) (2016) 113–134 . [10] G. Gan , M.K.-P. Ng , Subspace clustering with automatic feature grouping, Pat- tern Recognit. 48 (11) (2015) 3703–3713 . [11] W. Hong , J. Wright , K. Huang , Y.
Ma , Multiscale hybr id linear models for lossy image representation, IEEE Trans. Image Process. 15 (12) (2006) 3655–3671 . [12] J. Liu , Y.
Chen , J. Zhang , Z. Xu , Enhancing low-rank subspace clustering by man- ifold regularization, IEEE Trans. Image Process. 23 (9) (2014)
. [13] B. Li , Y. Zhang , Z. Lin , H. Lu , C.M.I. Center , Subspace clustering by mixture of gaussian regression, in: Proceedings of the IEEE Conference on Computer Vi- sion and Pattern Recognition, 2015, pp.
. [14] Z. Ma, A. Leijon, Bayesian estimation of beta mixture models with variational inference, IEEE Trans. Pattern Anal. Mach. Intell. 33 (11) (2011) 2160–2173, doi: 10.1109/TPAMI.2011.63 . [15] Z. Ma, A.E. Teschendorff, A. Leijon, Y. Qiao, H. Zhang, J. Guo, Variational Bayesian matrix factorization for bounded support data, IEEE Trans. Pattern Anal. Mach. Intell. 37
(4) (2015) 876–889, doi: 10.1109/TPAMI.2014.2353639 . [16] Z. Ma, J.H. Xue, A. Leijon, Z.H. Tan, Z. Yan g , J. Guo, Decorrelation of neutral vector variables: theory and applications, IEEE Trans. Neural Netw. Learn. Syst. PP (99) (2016) 1–15, doi: 10.1109/TNNL S.6445 . [17] V. Maz’ya , G. Schmidt , On approximate approximations using Gaussian kernels, IMA J. Numer. Anal. 16 (1) (1996) 13–29 . [18] D. Meng , F. Torre
, Robust matrix factorization with unknown noise, Proceed- ings of the International Conference on Computer Vision (2013) 1337–1344 . [19] Q. Zhao , D. Meng , Z. Xu , W.
Zuo , L. Zhang , Ro bust principal component analysis with complex noise, Proceedings of the International Conference on Machine Learning (2014) 55–63 . [20] X. Chen , Z. Han , Y. Wang , Q. Zhao , D. Meng , Y. Tang
, Robust tensor factoriza- tion with unknown noise, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 5213–5221 . [21] C. Lu , H. Min , Z. Zhao , L. Zhu , D. Huang , S. Yan
, Robust and efficient subspace segmentation via least squares regression, Proceedings of the European Con- ference on Computer Vision (2012) 347–360 . [22] R. Vidal , P. Favaro , Low ra nk subspace clustering (LRSC), Pattern Recognit. Lett. 43 (2014) 47–61 . [23] X. Cao , Y.
Chen , Q. Zhao , D. Meng , Y. Wang , D. Wang , Z. Xu , Low-rank matrix factorization under general mixture noise distributions, in: Proceedings of the International Conference on Computer Vision, 2015, pp. 1493–1501 . [24] T. Huang , H. Peng , K. Zhang , Model selection for Gaussian mixture models, Stat. Sin. 27 (1) (2017) 147–169 . [25] T.E . Boult , L.G. Brown , Factorization-based segmentation of motions, Proceed- ings of the IEEE Works ho p on Visual Motion (1991) 179–186 . [26] J.P. Costeira , T. Kanade , A multibody factorization method for independently moving objects, Int. J. Comput. Vision 29 (3) (1998) 159–179 . [27] C.W. Gear , Multibody grouping from motion images, Int. J. Comput. Vision 29 (2) (1998) 133–150 . [28] P. Ts eng
, Nearest q-flat to m points, J. Optim. Theory Appl. 105
(1) (20 0 0) 249–252 . [29] M.E. Tipping , C.M. Bishop , Mixtures of probabilistic principal component ana- lyzers, Neural Comput. 11
(2) (1999) 443–482 . [30] X. Lu , Y. Wa ng
, Graph-regularized low-rank representation for destrip- ing of hyperspectral images, IEEE Trans. Geosci. Remote Sens. 51 (7) (2013)
. [31] H. Lu , Z. Fu , X. Shu , Non-negative and sparse spectral clustering original re- search article, Pattern Recognit. 47 (1) (2014) 418–426 . [32] S.D. Babacan , S. Nakajima , M. Do , Probabilistic low-rank subspace clustering, Adv. Neural Inf. Process. Syst. 25 (2012)
. [33] S. Wang , X. Yuan , T. Yao
, J. Shen , Efficient subspace segmentation via quadratic programming, in: Proceedings of the 25th AAAI Conference on Arti- ficial Intelligence, 2011, pp. 519–524 . [34] R. Liu , Z. Lin , Z. Su , Learning Markov random walks for ro bust subspace clus- tering and estimation, IEEE Trans. Neural Netw. 59 (2014) 1–15 . [35] J. Feng , Z. Lin , H. Xu , S. Ya n , Robust subspace segmentation with block-diagonal prior, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp.
. [36] C. Lu , J. Tan g , M. Lin , L. Lin , S. Yan
, Z. Lin , Correntropy induce d L2 graph for ro- bust subspace clustering, Proceedings of the International Conference on Com- puter Vision (2013) 1801–1808 . [37] R. He , Y. Zhang , Z. Sun , Q. Yin , Robust subspace clustering with complex noise, IEEE Trans. Image Process. 24 (11) (2015)
. [38] P. Favaro , R. Vidal , A. Ravichandran , A closed form solution to robust subspace estimation and clustering, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2011, pp. 1801–1807 . [39] G. Schwarz , et al. , Estimating the dimension of a model, The Ann. Stat. 6 (2) (1978) 461–464 . [40] S. Wei, Z. Lin, Analysis and improvement of low rank representation for sub- space segmentation, arXiv: 1107. 156 1
(2011). [41] C.J. Wu , On the convergence properties of the EM algorithm, The Ann. Stat. 11
(1) (1983) 95–103 . [42] J. Yan , M. Pollefeys , A general framework for motion segmentation: indepen- dent, articulated, rigid, non-rigid, degenerate and non-degenerate, Proceedings of the European Conference on Computer Vision (2006) 94–106 . [43] A.M. Martinez, The AR face database, CVC Technical Report #24 (1998). [44] J. Wright , A.Y. Yan g , A. Ganesh , S.S. Sastry , Y. Ma , Robust face recognition via sparse representation, IEEE Trans. Pattern Anal. Mach. Intell. 31 (2) (2009) 210–227 . [45] R. Basri , D.W. Jacobs , Lambertian reflectance and linear subspaces, IEEE Trans. on Pattern Anal. Mach. Intell. 25 (2) (2003) 218–233 . [46] R. Tro n , R. Vidal , A benchmark for the comparison of 3-D motion segmentation algorithms, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2007, pp. 1–8 . [47] Y. Fu , J. Gao , D. Tien , Z. Lin , X. Hong , Tens or
LRR and sparse coding-based subspace clustering, IEEE Trans. Neural Netw. Learn. Syst. 27 (10) (2016) 2120–2133 . Jing Yao
received the B.Sc. degree from Northwestern University, Xi’an, China, in 2014. He is currently pursu- ing the Ph.D. degree in Xi’an Jiaotong University. His cur- rent research interests include low-rank matrix factoriza- tion and hyperspectral image analysis. Xiangyong Cao received the B.Sc. degree from Xi’an Jiao- tong University, Xi’an, China, in 2012, where he is cur- rently pursuing the Ph.D. degree. His current research interests include low-rank matrix analysis, and Bayesian method for machine learning. Qian Zhao received the B.Sc. and Ph.D. degrees from Xi’an Jiaotong University, Xi’an, China, in 2009 and 2015, re- spectively. He was a Visiting Scholar with Carnegie Mel- lon University, Pittsburgh, PA,
USA, from 2013 to 2014. He is currently a Lecturer with the School of Mathe- matics and Statistics, Xi’an Jiaotong University. His cur- rent research interests include low-matrix/ tensor analy- sis, Bayesian modeling, and self-paced learning. Deyu Meng received the B.Sc., M.Sc., and Ph.D. degrees from Xi’an Jiaotong University, Xi’an, China, in 2001, 2004, and 2008, respectively. He was a Visiting Scholar with Carnegie Mellon University, Pittsburgh, PA , USA , from 2012 to 2014. He is currently an Associate Professor with the Institute for Information and System Sciences, School of Mathematics and Statistics, Xi’an Jiaotong University. His current research interests include principal compo- nent analysis, nonlinear dimensionality reduction, feature extraction and selection, compressed sensing, and sparse machine learning methods. Zongben Xu received the Ph.D. degree in mathemat- ics from Xi’an Jiaotong University, Xi’an, China, in 198 7. He currently serves as the Academician of the Chinese Academy of Sciences, the Chief Scientist of the National Basic Research Program of China (973 Project), and the Director of the Institute for Information and System Sci- ences with Xi’an Jiaotong University. His current research interests include nonlinear functional analysis and intel- ligent information processing. Prof. Xu was a recipient of the National Natural Science Award of China in 2007 and the winner of the CSIAM Su Buchin Applied Mathemat- ics Prize in 2008. He delivered a talk at the International Congress of Mathematicians in 2010. Please cite this article as: J. Yao
et al., Robust subspace clustering via penalized mixture of Gaussians, Neurocomputing (2017), http://dx.doi.org/10.1016/j.neucom.2017.05.102
Conference PaperFull-text availableJun 2015Conference PaperJun 2016ArticleOct 2016ArticleJan 2014ArticleApr 2016Conference PaperJun 2014ArticleApr 2016ArticleJan 2015ArticleJul 2015ArticleJan 1996Show moreProject[...]Project[...]Project[...]Project[...]ArticleJanuary 2016Many computer vision problems can be posed as learning a low-dimensional
subspace from high dimensional data. The low rank matrix factorization (LRMF)
represents a commonly utilized subspace learning strategy. Most of the current
LRMF techniques are constructed on the optimization problems using L1-norm and
L2-norm losses, which mainly deal with Laplacian and Gaussian noises,
respectively. To... [Show full abstract]ArticleOctober 2016 · Many computer vision problems can be posed as learning a low-dimensional subspace from high-dimensional data. The low rank matrix factorization (LRMF) represents a commonly utilized subspace learning strategy. Most of the current LRMF techniques are constructed on the optimization problems using L-1-norm and L-2-norm losses, which mainly deal with the Laplace and Gaussian noises, respectively.... [Show full abstract]ArticleJanuary 2017 · Hyperspectral image (HSI) denoising has been attracting much research attention in remote sensing area due to its importance in improving the HSI qualities. The existing HSI denoising methods mainly focus on specific spectral and spatial prior knowledge in HSIs, and share a common underlying assumption that the embedded noise in HSI is independent and identically distributed (i.i.d.). In real... [Show full abstract]ArticleNovember 2016 · Hyperspectral image (HSI) classification is one of the fundamental tasks in HSI analysis. Recently, many approaches have been extensively studied to improve the classification performance, among which integrating the spatial information underlying HSIs is a simple yet effective way. However, most of the current approaches haven't fully exploited the spatial information prior. They usually... [Show full abstract]}

我要回帖

更多关于 打call是什么意思 的文章

更多推荐

版权声明:文章内容来源于网络,版权归原作者所有,如有侵权请点击这里与我们联系,我们将及时删除。

点击添加站长微信