Jump to content

Margin-infused relaxed algorithm

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Jusdafax (talk | contribs) at 12:48, 8 July 2010 (Reverted addition of dubious unsourced content (HG)). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Margin Infused Relaxed Algorithm (MIRA)[1] is a machine learning algorithm, an online algorithm for multiclass classification problems. It is designed to learn a set of parameters (vector or matrix) by processing all the given training examples one-by-one and updating the parameters according to each training example, so that the current training example is classified correctly with a margin against incorrect classifications at least as large as their loss[2]. The change of the parameters is kept as small as possible.

The flow of the algorithm[3][4] looks as follows:

Algorithm MIRA
  Input: Training examples 
  Output: Set of parameters 
  , 
  for  to 
    for  to 
       update  according to 
      
    end for
  end for
  return 
  • "←" denotes assignment. For instance, "largestitem" means that the value of largest changes to the value of item.
  • "return" terminates the algorithm and outputs the following value.

The update step is then formalized as an optimization problem: Find , so that .

References

  1. ^ Crammer, K., Singer, Y. (2003): Ultraconservative Online Algorithms for Multiclass Problems. In: Journal of Machine Learning Research 3, 951-991.
  2. ^ McDonald, R. et al (2005): Online Large-Margin Training of Dependency Parsers. In: Proceedings of the 43rd Annual Meeting of the ACL, pp. 91-98.
  3. ^ Wanatabe, T. et al (2007): Online Large Margin Training for Statistical Machine Translation. In: Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 764–773.
  4. ^ Bohnet, B. (2009): Efficient Parsing of Syntactic and Semantic Dependency Structures. Proceedings of Conference on Natural Language Learning (CoNLL), Boulder, 67-72.