Abstract
We propose a new algorithm for minimizing regularized empirical loss: Stochastic Dual Newton Ascent (SDNA). Our method is dual in nature: in each iteration we update a random subset of the dual variables. However, unlike existing methods such as stochastic dual coordinate ascent. SDNA is capable of utilizing all local curvature information contained in the examples, which leads to striking improvements in both theory and practice - sometimes by orders of magnitude. In the special case when an L2-regularizer is used in the primal, the dual problem is a concave quadratic maximization problem plus a separable term. In this regime, SDNA in each step solves a proximal subproblem involving a random principal submatrix of the Hessian of the quadratic function; whence the name of the method.
Original language | English (US) |
---|---|
Title of host publication | 33rd International Conference on Machine Learning, ICML 2016 |
Editors | Kilian Q. Weinberger, Maria Florina Balcan |
Publisher | International Machine Learning Society (IMLS) |
Pages | 2707-2725 |
Number of pages | 19 |
ISBN (Electronic) | 9781510829008 |
State | Published - 2016 |
Externally published | Yes |
Event | 33rd International Conference on Machine Learning, ICML 2016 - New York City, United States Duration: Jun 19 2016 → Jun 24 2016 |
Publication series
Name | 33rd International Conference on Machine Learning, ICML 2016 |
---|---|
Volume | 4 |
Other
Other | 33rd International Conference on Machine Learning, ICML 2016 |
---|---|
Country/Territory | United States |
City | New York City |
Period | 06/19/16 → 06/24/16 |
Bibliographical note
Publisher Copyright:© 2016 by the author(s).
ASJC Scopus subject areas
- Artificial Intelligence
- Software
- Computer Networks and Communications