Random feature
Random features (RF) are a technique used in machine learning to approximate kernel methods, introduced by Ali Rahimi and Ben Recht in their 2007 paper "Random Features for Large-Scale Kernel Machines".[1]
RF uses a Monte Carlo approximation to kernel functions by randomly sampled feature maps. It is used for datasets that are too large for traditional kernel methods like support vector machine, kernel ridge regression, and gaussian process.
Mathematics
Given a feature map , where is a Hilbert space, the kernel trick replaces inner products in high-dimensional feature space by a kernel function that is positive-definite.
Kernel methods perform linear operations in high-dimensional space by manipulating the kernel matrix instead: where is the number of data points.
The problem with kernel methods is that the kernel matrix has size . This becomes computationally infeasible when reaches the order of 1 million.
The random kernel method replaces the kernel function by an inner product in low-dimensional feature space : where is a randomly sampled feature map .
This converts kernel linear regression into linear regression in feature space, kernel SVM into feature SVM, etc. These methods no longer involve matrices of size , but only random feature matrices of size .
Examples
Radial basis function kernel
The Radial basis function kernel can be approximated by random Fourier transformation of the kernel:where are independent samples from the normal distribution .
Theorem: (unbiased esimation)
Proof: It suffices to prove the case of . Use the trigonometric identity , the spherical symmetry of gaussian distribution, then evaluate the integral
Theorem: (convergence) As the number of random features increases, the approximation converges to the true kernel with high probability.
Theorem: (variance bound) . (Appendix A.2[2]).
Orthogonal random features
Orthogonal random features[3] uses a random orthogonal matrix instead of a random Fourier matrix.
See also
References
- ^ Rahimi, Ali; Recht, Benjamin (2007). "Random Features for Large-Scale Kernel Machines". Advances in Neural Information Processing Systems. 20.
- ^ Peng, Hao; Pappas, Nikolaos; Yogatama, Dani; Schwartz, Roy; Smith, Noah A.; Kong, Lingpeng (2021-03-19). "Random Feature Attention". arXiv:2103.02143 [cs.CL].
- ^ Yu, Felix Xinnan X; Suresh, Ananda Theertha; Choromanski, Krzysztof M; Holtmann-Rice, Daniel N; Kumar, Sanjiv (2016). "Orthogonal Random Features". Advances in Neural Information Processing Systems. 29. Curran Associates, Inc.