Distributed optimization algorithms are essential for training machine
learning models on very large-scale datasets. However, they often suffer from
communication bottlenecks. Confronting this issue, a communication-efficient
primal-dual coordinate ascent framework (CoCoA) and its improved variant CoCoA+
have been proposed, achieving a convergence rate of $O(1/t)$ for
solving empirical risk minimization problems with Lipschitz continuous losses.
In this paper, an accelerated variant of CoCoA+ is proposed and shown to
possess a convergence rate of $O(1/t^2)$ in terms of reducing
suboptimality. The analysis of this rate is also notable in that the
convergence rate bounds involve constants that, except in extreme cases, are
significantly reduced compared to those previously provided for CoCoA+. The
results of numerical experiments are provided to show that acceleration can
lead to significant performance gains.
Beschreibung
An Accelerated Communication-Efficient Primal-Dual Optimization Framework for Structured Machine Learning
%0 Generic
%1 ma2017accelerated
%A Ma, Chenxin
%A Jaggi, Martin
%A Curtis, Frank E.
%A Srebro, Nathan
%A Takáč, Martin
%D 2017
%K deep dl large-scale networks neural
%T An Accelerated Communication-Efficient Primal-Dual Optimization
Framework for Structured Machine Learning
%U http://arxiv.org/abs/1711.05305
%X Distributed optimization algorithms are essential for training machine
learning models on very large-scale datasets. However, they often suffer from
communication bottlenecks. Confronting this issue, a communication-efficient
primal-dual coordinate ascent framework (CoCoA) and its improved variant CoCoA+
have been proposed, achieving a convergence rate of $O(1/t)$ for
solving empirical risk minimization problems with Lipschitz continuous losses.
In this paper, an accelerated variant of CoCoA+ is proposed and shown to
possess a convergence rate of $O(1/t^2)$ in terms of reducing
suboptimality. The analysis of this rate is also notable in that the
convergence rate bounds involve constants that, except in extreme cases, are
significantly reduced compared to those previously provided for CoCoA+. The
results of numerical experiments are provided to show that acceleration can
lead to significant performance gains.
@misc{ma2017accelerated,
abstract = {Distributed optimization algorithms are essential for training machine
learning models on very large-scale datasets. However, they often suffer from
communication bottlenecks. Confronting this issue, a communication-efficient
primal-dual coordinate ascent framework (CoCoA) and its improved variant CoCoA+
have been proposed, achieving a convergence rate of $\mathcal{O}(1/t)$ for
solving empirical risk minimization problems with Lipschitz continuous losses.
In this paper, an accelerated variant of CoCoA+ is proposed and shown to
possess a convergence rate of $\mathcal{O}(1/t^2)$ in terms of reducing
suboptimality. The analysis of this rate is also notable in that the
convergence rate bounds involve constants that, except in extreme cases, are
significantly reduced compared to those previously provided for CoCoA+. The
results of numerical experiments are provided to show that acceleration can
lead to significant performance gains.},
added-at = {2019-06-04T19:05:20.000+0200},
author = {Ma, Chenxin and Jaggi, Martin and Curtis, Frank E. and Srebro, Nathan and Takáč, Martin},
biburl = {https://www.bibsonomy.org/bibtex/235307d3d3cbff45b9b4ea10254cb48d6/alrigazzi},
description = {An Accelerated Communication-Efficient Primal-Dual Optimization Framework for Structured Machine Learning},
interhash = {1f0bad92efdca48a6e8e7b1ebfe752e6},
intrahash = {35307d3d3cbff45b9b4ea10254cb48d6},
keywords = {deep dl large-scale networks neural},
note = {cite arxiv:1711.05305},
timestamp = {2019-06-04T19:05:20.000+0200},
title = {An Accelerated Communication-Efficient Primal-Dual Optimization
Framework for Structured Machine Learning},
url = {http://arxiv.org/abs/1711.05305},
year = 2017
}