User:Numerical one/sandbox
In numerical optimization, the Broyden–Fletcher–Goldfarb–Shanno (BFGS) algorithm is an iterative method for solving unconstrained nonlinear optimization problems.
The BFGS method belongs to quasi-Newton methods, a class of hill-climbing optimization techniques that seek a stationary point of a (preferably twice-continuously differentiable) function using only gradient evaluations (or approximate gradient evaluations). For such problems, a necessary condition for optimality at a point is that the gradient at f at is zero, i.e., . Under certain conditions, the BFGS method exhibits superlinear convergence. (citation needed) For large problems, limited-memory BFGS (L-BFGS) is commonly used. The BFGS-B variant handles simple box constraints.
The BFGS method is one of the most popular types of quasi-Newton methods. In the case of large scale optimization (i.e., when the number of variables is greater than, for example, 10,000), limited-memory BFGS (L-BFGS) may be used. Other variants include BFGS-B, which handles simple box constraints. citation needed!
The algorithm is named after Charles George Broyden, Roger Fletcher, Donald Goldfarb and David Shanno. (I think we can say more about how these authors found this!)
Rationale
[edit]The BFGS algorithm is a quasi-Newton method for unconstrained optimization where the Hessian is approximated using the BFGS update formula. The BFGS algorithm is used to solve problems of the form
minimize subject to where This algorithm is an iterative method where the search direction is computed as the solution of minimizing a quadratic approximation to at the current iterate i.e., where and is the BFGS matrix that approximates the Hessian Typically, a line search is then used to compute so that the next iterate is defined as
Derivation
[edit]Like all traditional quasi-Newton matrices, the BFGS matrix satisfies the so-called secant or quasi-Newton condition: where and The BFGS update is a symmetric update that is derived by considering a rank-two change to the previous approximate Hessian, i.e., where and are vectors. One can show that choosing and then and in order that the quasi-Newton condition holds. Thus, the BFGS update is
It can be shown that if the initial is positive definite and for all then the sequence of matrices will be positive definite. A Wolfe line search is typically used in conjunction with the BFGS algorithm since it guarantees (citation needed!)