Jump to content

User:Aali9520/Tensor Completion

From Wikipedia, the free encyclopedia

In mathematics, tensor completion (also known as tensor recovery) is the recovery of a tensor when a subset of its entries are missing or distorted. Tensors are commonly used to represent multidimensional data and the problem of estimating them from data is a generalization of matrix completion to higher dimensions. Tensor completion is also closely related to higher-order singular value decomposition. Analogous to the matrix completion problem, structural assumptions about the tensor are usually made, with the most common being low rank. The remainder of this article focuses specifically on low rank tensor completion.

Tensors

[edit]

A tensor is a multidimensional array. Formally, let be a tensor of order and size . The entry of is denoted . As a running example (taken from [1]), let be the order 3 tensor whose frontal slices are

Fibers are the higher-order analogue of rows and columns in matrices, defined by fixing all but one index. The mode-n fibers are all of the fibers obtained by fixing all indices except that of dimension n. For example, the mode-1 fibers and mode-2 fibers of a matrix are respectively its columns and rows. In the running example, there are twelve mode-3 fibers of , each of length two with corresponding elements taken from the two frontal slices.

The mode-n matricization (also known as unfolding or flattening) of , denoted , is the matrix whose columns are the mode-n fibers. The specific ordering of the columns will not matter here. The three mode-n unfoldings of our running example are

The n-rank of , denoted , is defined to be the matrix rank of the mode-n matricization:

This concept of tensor rank is closely related to the Tucker decomposition.

Low rank tensor completion

[edit]

Let be the true tensor to be recovered. It is assumed that the n-ranks are "small," i.e. is small compared to for some or all n.

Completing missing entries

[edit]

A standard problem is to observe some entries of and fill in the remaining entries. Let be the set of indices observed. Then the goal is to recover the tensor of minimum rank among all tensors that match the data. This is formulated as

.

Unfortunately, this problem is difficult to solve as the function is not convex. In fact, even in the special case that is a matrix, the problem is NP-hard. [2]

In matrix completion, a common approach is to replace the rank in the objective function with the 1-schatten norm (also called nuclear norm), denoted . This gives a convex problem that is tractable and enjoys a number of recovery guarantees. One such guarantee is the celebrated result of Emmanuel Candès and Benjamin Recht [3]: given a matrix of rank satisfying certain incoherence properties, if entries are chosen uniformly at random and observed, then the matrix can be exactly recovered with high probability from nuclear norm minimization as long as .

The same idea can be extended to the tensor completion problem above, i.e. the following convex problem is solved:

General tensor recovery

[edit]

The information about may be more general than above. Let be a linear map, and let be a possibly noisy observation of . Then the problem is for some appropriate choice of :

Recovery guarantees

[edit]

Finding recovery guarantees for nuclear norm minimization in tensor completion that are analogous to existing guarantees in matrix completion remains an open problem. One result [4] for the case of noisy observations states that for an order tensor that is balanced, i.e. of size and for all , if each entry is distorted by an independent Gaussian variable with mean zero and variance , then nuclear norm minimization recovers the original tensor with high probability as long as . This can be extended to unbalanced tensors.

Algorithms

[edit]

There is a large amount of research into solution techniques for the low rank tensor recovery formulation above. For convenience, the unconstrained, penalized formulation is usually considered:

The problem falls into the class of convex optimization problems, for which a variety of solution methods exist. However, some methods may be inappropriate for this particular objective function given the sum of many functions and the non-differentiability of the nuclear norm. Furthermore, tensor data is often very large in practice, making standard methods intractable.

Proximal gradient methods, designed for optimizing the sum of convex, possibly non-differentiable functions, are a natural candidate. In particular, this problem fits into the area of proximal gradient methods for learning. Some solution approaches that have had computational success are the Alternating-direction method of multipliers and the Douglas-Rachford splitting technique [5], though this still remains an active area of research.

Applications

[edit]

One major application of tensor completion is the recovery of visual data. Digital two-dimensional monochromatic images can be represented as a matrix, where each entry is the intensity of the corresponding pixel, and matrix completion techniques can be used to recover images with missing or distorted entries. Similarly, tensors of order three can be used to represent three-dimensional images, where each entry is the intensity of the corresponding voxel. Three-dimensional images often arise in medical and scientific settings. For example, magnetic resonance imaging produces three-dimensional images, and tensor completion can be used when observations are missing or distorted. There are heuristic methods for applying matrix completion to three-dimensional images, including completing the entries of one of the matricizations, or completing two-dimensional slices of the image separately, but tensor completion has been shown experimentally to outperform these heuristics. [6]

In addition, tensors can represent a wide variety of other visual data. Monochromatic videos can be represented as tensors of order 3, where two-dimensional images are stacked along the time axis. Polychromatic data can also be represented by adding an additional dimension to contain the multiple values needed to represent colors.

There are many other applications of tensors in which low rank completion is used, including

See also

[edit]

References

[edit]
  1. ^ Kolda, Tamara G.; Bader, Brett W. (2009). "Tensor Decompositions and Applications". SIAM Review. 51 (3): 455–500. doi:10.1137/07070111X.
  2. ^ Vandenberghe, Lieven; Boyd, Stephen (March 1996). "Semidefinite Programming" (PDF). SIAM Review. 38 (1): 49–95.
  3. ^ Candes, Emmanuel J.; Recht, Benjamin (December 2009). "Exact Matrix Completion via Convex Optimization". Foundations of Computational Mathematics. 9 (6): 717–772.
  4. ^ Tomioka, Ryota; Suzuki, Taiji; Hayashi, Kohei; Kashima, Hisashi (2011). "Statistical Performance of Convex Tensor Decomposition" (PDF). Advances in Neural Information Processing Systems. 24: 972–980.
  5. ^ Gandy, Silvia; Recht, Benjamin; Yamada, Isao (2011). "Tensor completion and low-n-rank tensor recovery via convex optimization" (PDF). Inverse Problems. 27 (2).
  6. ^ Liu, Ji; Musialski, Przemyslaw; Wonka, Peter; Ye, Jieping (2012). "Tensor Completion for Estimating Missing Values in Visual Data". IEEE Transactions on Pattern Analysis and Machine Intelligence. 35 (1): 208–220. doi:10.1109/TPAMI.2012.39.
  7. ^ Mesgarani, N.; Slaney, M.; Shamma, S.A. (May 2006). "Discrimination of speech from nonspeech based on multiscale spectro-temporal Modulations". IEEE Transactions on Audio, Speech, and Language Processing. 14 (3): 920–930. doi:10.1109/TSA.2005.858055.
  8. ^ Cohen, Shay B.; Collins, Michael (2012). "Tensor Decomposition for Fast Parsing with Latent-Variable PCFGs" (PDF). Advances in Neural Information Processing Systems. 25.
  9. ^ Nion, D.; Sidiropoulos, N.D. (2010). "Tensor Algebra and Multidimensional Harmonic Retrieval in Signal Processing for MIMO Radar". IEEE Transactions on Signal Processing. 58 (11): 5693–5705. doi:10.1109/TSP.2010.2058802.
  10. ^ Anandkumar, Anima; Ge, Rong; Hsu, Daniel; Kakade, Sham M.; Telgarsky, Matus (2014). "Tensor Decompositions for Learning Latent Variable Models" (PDF). Journal of Machine Learning Research. 15: 2773–2832.
  11. ^ Romera-Paredes, Bernardino; Aung, Min Hane; Bianchi-Berthouze, Nadia; Pontil, Massimiliano (2013). "Multilinear Multitask Learning" (PDF). Proceedings of the 30th International Conference on Machine Learning. 28.