Prior knowledge for pattern recognition
free lunch theorem states that all search algorithms have the same average performance over all problems, and thus implies that to gain in performance on a certain application one must use a specialized algorithm that includes some prior knowledge about the problem.
The different types of prior knowledge encountered in pattern recognition are now regrouped under two main categories: class-invariance and knowledge on the data.
Class-invariance
A very common type of prior knowledge in pattern recognition is the invariance of the class (or the output of the classifier) to a transformation of the input pattern. This type of knowledge is referred to as transformation-invariance. The mostly used transformations used in image recognition are:
Incorporating the invariance to a transformation parametrized in into a classifier of output for an input pattern corresponds to enforce the equality
Local invariance can also be considered for a transformation centered at , so that , by the constraint
It must be noted that in these Equations can be either the decision function of the classifier or its real-valued output.
Another approach is to consider the class-invariance with respect to a "domain of the input space" instead of a transformation. In this case, the problem becomes finding so that
where is the membership class of the region of the input space.
A different type of class-invariance found in pattern recognition is the permutation-invariance, i.e. invariance of the class to a permutation of elements in a structured input. A typical application of this type of prior knowledge is a classifier invariant to permutations of rows in matrix inputs.
Knowledge on the data
Other forms of prior knowledge than class-invariance concern the data more specifically and are thus of particular interest for real-world applications. The three particular cases that most often occur when gathering data are:
- Unlabeled samples are available with supposed class-memberships;
- Imbalance of the training set due to a high proportion of samples of a class;
- Quality of the data may vary from a sample to another.
Prior knowledge on these can enhance the quality of the recognition if included in the learning. Moreover, not taking into account the poor quality of some data or a large imbalance between the classes can mislead the decision of a classifier.
References
- [Scholkopf02], B. Scholkopf and A. Smola, "Learning with Kernels", MIT Press 2002.
- [Krupka07], E. Krupka and N. Tishby, "Incorporating Prior Knowledge on Features into Learning", Eleventh International Conference on Artificial Intelligence and Statistics (AISTATS 07)