Jump to content

User:Makaylapark/Hyperparameter (machine learning)

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Makaylapark (talk | contribs) at 04:59, 3 May 2022 (add citation). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Hyperparameters can be classified as model hyperparameters, that cannot be inferred while fitting the machine to the training set because they refer to the model selection task, or algorithm hyperparameters, that in principle have no influence on the performance of the model but affect the speed and quality of the learning process. An example of a model hyperparameter is the topology and size of a neural network. Examples of algorithm hyperparameters are learning rate and batch size as well as mini-batch size. A batch size can refer to the full sample whereas a mini-batch is a smaller sample set.

Different model training algorithms require different hyperparameters, some simple algorithms (such as ordinary least squares regression) require none. Given these hyperparameters, the training algorithm learns the parameters from the data. For instance, LASSO is an algorithm that adds a regularization hyperparameter to ordinary least squares regression, which has to be set before estimating the parameters through the training algorithm.[1]

Article Draft

Lead

Article body

References

  1. ^ Yang, Li; Shami, Abdallah (2020-11-20). "On hyperparameter optimization of machine learning algorithms: Theory and practice". Neurocomputing. 415: 295–316. doi:10.1016/j.neucom.2020.07.061. ISSN 0925-2312.