Jump to content

L1-norm principal component analysis

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Cisumlacissalc (talk | contribs) at 20:27, 21 May 2018 (Approximate/Efficient Solvers). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

L1-norm principal component analysis (L1-PCA) is a method for robust multivariate data analysis.[1] L1-PCA is often preferred over standard principal component analysis (PCA) when the analyzed data may contain outliers (faulty points, irregular corruptions).[2][3][4]

PCA seeks a collection of orthogonal directions (principal components) that define a subspace wherein data representation is maximized.[5][6][7] Standard PCA quantifies data representation as the aggregate L2-norm of the data point projections into the subspace, or equivalently the aggregate Euclidean distance of the original points from their subspace-projected representations. Therefore, standard PCA is also referred to L2-PCA, mostly to be distinguished from L1-PCA.[8] In PCA and L1-PCA, the number of principal components (PCs) is lower than the rank of the analyzed matrix, which coincides with the dimensionality of space defined by the original data points. Therefore, PCA is commonly employed for dimensionality reduction, e.g., for the purpose of denoising, or compression. Among the advantages of PCA that have contributed to its high popularity are its low-cost implementation by means of singular-value decomposition (SVD)[9] and its quality approximation of the maximum-variance subspace of the data source, for certain data distributions, such as multivariate Normal, when it operates on sufficiently many nominal data points.

However, in modern big data sets, data often include grossly corrupted, faulty points, commonly referred to as outliers.[10] Regretfully, standard PCA is known to be very sensitive against outliers, even when they appear as a small fraction of the processed data.[11] The reason is the L2-norm formulation of PCA places squared emphasis to the magnitude of each coordinate of each data point, ultimately benefiting peripheral points, such as outliers. On the other hand, following an L1-norm formulation, L1-PCA places linear emphasis on the coordinates of each data point, thus counteracting and effectively restraining outliers.[12]

Formulation

Consider matrix consisting of -dimensional data points. Define . For integer such that , L1-PCA is formulated as:[1]

For , (2) simplifies to finding the L1-norm principal component (L1-PC) of as

In (1)-(2), L1-norm returns the sum of the absolute entries of its argument and L2-norm returns the sum of the squared entries of its argument. If one substitutes in (2) by the Frobenius/L2-norm , then the problem becomes standard PCA and it is solved by the matrix that contains the dominant singular vectors of (i.e., the singular vectors that correspond to the highest singular values).

The maximization metric in (2) can be expanded as

Solution

For any matrix with , define as the nearest (in the L2-norm sense) matrix to that has orthonormal columns. That is, define

Procrustes Theorem[13][14] states that if has SVD , then .

Markopoulos et al.[1] showed that, if is the exact solution to the binary nuclear-norm maximization (BNM)

then

is the exact solution to L1-PCA in (2). The nuclear-norm in (2) returns the summation of the singular values of its matrix argument and can be calculated by means of standard SVD. Moreover, it holds that, given the solution to L1-PCA, , the solution to BNM can be obtained as

where returns the -sign matrix of its matrix argument (with no loss of generality, we can consider ). In addition, it follows that . Clearly, BNM in (5) is a combinatorial problem over antipodal binary variables. Therefore, its exact solution can be found through exhaustive evaluation of all elements of its feasibility set, with asymptotic cost . Therefore, L1-PCA can also be solved, through BNM, with cost (exponential in the product of the number of data points with the number of the sought-after components).

For the special case of (single L1-PC of ), BNM takes the binary-quadratic-maximization (BQM) form

The transition from (5) to (7) for holds true, since the unique singular value of is equal to , for every . Then, if is the solution to BQM in (7), it holds that

is the exact L1-PC of , as defined in (1). In addition, it holds that and .

Algorithms

Exact Solution

As shown above, the exact solution to L1-PCA can derive by the following two-step process:

1. Solve problem in (5) to obtain .
2. Apply SVD on  to obtain .

BNM in (5) can be solved by exhaustive search of the feasibility set with cost . Also, it can be solved with cost , when is considered a constant with respect to .[1][15]


Approximate/Efficient Solvers

In 2008, Kwak[12] proposed an iterative algorithm for the solution of L1-PCA, for . This iterative method was later generalized for components.[16] Another algorithm was proposed for solving BNM in (5) by means of semi-definite programming (SDP). Most recently, L1-PCA in (2) and BNM in (5) were solved efficiently by means of bit-flipping iterations (L1-BF algorithm).

L1-BF Algorithm

 1  function L1BF(, ):
 2      Initialize  and 
 3      Set  and  
 4      Until termination (or  iterations)
 5          , 
 6          For 
 7              , 
 8                              // flip bit
 9                             // calculated by SVD or faster (see[17])
10              if 
11                  , 
12                  
13              end
14              if                     // no bit was flipped
15                  if 
16                      terminate
17                  else
18                      

References

  1. ^ a b c d Markopoulos, Panos P.; Karystinos, George N.; Pados, Dimitris A. (October 2014). "Optimal Algorithms for L1-subspace Signal Processing". IEEE Transactions on Signal Processing. 62 (19): 5046–5058. doi:10.1109/TSP.2014.2338077.
  2. ^ Barrodale, I. (1968). "L1 Approximation and the Analysis of Data". Applied Statistics. 17 (1): 51. doi:10.2307/2985267.
  3. ^ Barnett, Vic; Lewis, Toby (1994). Outliers in statistical data (3. ed. ed.). Chichester [u.a.]: Wiley. ISBN 0471930946. {{cite book}}: |edition= has extra text (help)
  4. ^ Kanade, T.; Ke, Qifa (June 2005). "Robust L1 Norm Factorization in the Presence of Outliers and Missing Data by Alternative Convex Programming". 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05). IEEE. doi:10.1109/CVPR.2005.309.
  5. ^ Jolliffe, I.T. (2004). Principal component analysis (2nd ed. ed.). New York: Springer. ISBN 0387954422. {{cite book}}: |edition= has extra text (help)
  6. ^ Bishop, Christopher M. (2007). Pattern recognition and machine learning (Corr. printing. ed.). New York: Springer. ISBN 978-0-387-31073-2.
  7. ^ Pearson, Karl (8 June 2010). "On Lines and Planes of Closest Fit to Systems of Points in Space". The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science. 2 (11): 559–572. doi:10.1080/14786440109462720.
  8. ^ Markopoulos, Panos P.; Kundu, Sandipan; Chamadia, Shubham; Pados, Dimitris A. (15 August 2017). "Efficient L1-Norm Principal-Component Analysis via Bit Flipping". IEEE Transactions on Signal Processing. 65 (16): 4252–4264. doi:10.1109/TSP.2017.2708023.
  9. ^ Golub, Gene H. (April 1973). "Some Modified Matrix Eigenvalue Problems". SIAM Review. 15 (2): 318–334. doi:10.1137/1015032.
  10. ^ Barnett, Vic; Lewis, Toby (1994). Outliers in statistical data (3. ed. ed.). Chichester [u.a.]: Wiley. ISBN 0471930946. {{cite book}}: |edition= has extra text (help)
  11. ^ Candès, Emmanuel J.; Li, Xiaodong; Ma, Yi; Wright, John (1 May 2011). "Robust principal component analysis?". Journal of the ACM. 58 (3): 1–37. doi:10.1145/1970392.1970395.
  12. ^ a b Kwak, N. (September 2008). "Principal Component Analysis Based on L1-Norm Maximization". IEEE Transactions on Pattern Analysis and Machine Intelligence. 30 (9): 1672–1680. doi:10.1109/TPAMI.2008.114.
  13. ^ Eldén, Lars; Park, Haesun (1 June 1999). "A Procrustes problem on the Stiefel manifold". Numerische Mathematik. 82 (4): 599–619. doi:10.1007/s002110050432.
  14. ^ Schönemann, Peter H. (March 1966). "A generalized solution of the orthogonal procrustes problem". Psychometrika. 31 (1): 1–10. doi:10.1007/BF02289451.
  15. ^ Markopoulos, PP; Kundu, S; Chamadia, S; Tsagkarakis, N; Pados, DA (2018). "Outlier-Resistant Data Processing with L1-Norm Principal Component Analysis". Advances in Principal Component Analysis (Springer, Singapore). doi:10.1007/978-981-10-6704-4_6.
  16. ^ Nie, F; Huang, H; Ding, C; Luo, Dijun; Wang, H (July 2011). "Robust principal component analysis with non-greedy l1-norm maximization". 22nd International Joint Conference on Artificial Intelligence: 1433–1438.
  17. ^ mark2017