Invention Grant
US07933847B2 Limited-memory quasi-newton optimization algorithm for L1-regularized objectives 有权
L1规范化目标的有限存储准牛顿优化算法

Limited-memory quasi-newton optimization algorithm for L1-regularized objectives
Abstract:
An algorithm that employs modified methods developed for optimizing differential functions but which can also handle the special non-differentiabilities that occur with the L1-regularization. The algorithm is a modification of the L-BFGS (limited-memory Broyden-Fletcher-Goldfarb-Shanno) quasi-Newton algorithm, but which can now handle the discontinuity of the gradient using a procedure that chooses a search direction at each iteration and modifies the line search procedure. The algorithm includes an iterative optimization procedure where each iteration approximately minimizes the objective over a constrained region of the space on which the objective is differentiable (in the case of L1-regularization, a given orthant), models the second-order behavior of the objective by considering the loss component alone, using a “line-search” at each iteration that projects search points back onto the chosen orthant, and determines when to stop the line search.
Information query
Patent Agency Ranking
0/0