We address the issue of speeding up the training of convolutional neural
networks by studying a distributed method adapted to stochastic gradient
descent. Our parallel optimization setup uses several threads, each applying
individual gradient descents on a local variable. We propose a new way of
sharing information between different threads based on gossip algorithms that
show good consensus convergence properties. Our method called GoSGD has the
advantage to be fully asynchronous and decentralized.
Description
GoSGD: Distributed Optimization for Deep Learning with Gossip Exchange
%0 Generic
%1 blot2018gosgd
%A Blot, Michael
%A Picard, David
%A Cord, Matthieu
%D 2018
%K ASGD distributed to_read
%T GoSGD: Distributed Optimization for Deep Learning with Gossip Exchange
%U http://arxiv.org/abs/1804.01852
%X We address the issue of speeding up the training of convolutional neural
networks by studying a distributed method adapted to stochastic gradient
descent. Our parallel optimization setup uses several threads, each applying
individual gradient descents on a local variable. We propose a new way of
sharing information between different threads based on gossip algorithms that
show good consensus convergence properties. Our method called GoSGD has the
advantage to be fully asynchronous and decentralized.
@misc{blot2018gosgd,
abstract = {We address the issue of speeding up the training of convolutional neural
networks by studying a distributed method adapted to stochastic gradient
descent. Our parallel optimization setup uses several threads, each applying
individual gradient descents on a local variable. We propose a new way of
sharing information between different threads based on gossip algorithms that
show good consensus convergence properties. Our method called GoSGD has the
advantage to be fully asynchronous and decentralized.},
added-at = {2018-04-08T17:08:19.000+0200},
author = {Blot, Michael and Picard, David and Cord, Matthieu},
biburl = {https://www.bibsonomy.org/bibtex/2e1bb53b02bcdca7bd8723aa6b338eaf6/jk_itwm},
description = {GoSGD: Distributed Optimization for Deep Learning with Gossip Exchange},
interhash = {bbd2cc154b2c50642f5a3b9b7b6d17ee},
intrahash = {e1bb53b02bcdca7bd8723aa6b338eaf6},
keywords = {ASGD distributed to_read},
note = {cite arxiv:1804.01852},
timestamp = {2018-04-08T17:08:19.000+0200},
title = {GoSGD: Distributed Optimization for Deep Learning with Gossip Exchange},
url = {http://arxiv.org/abs/1804.01852},
year = 2018
}